Enable Ceph’s RADOS Gateway (RGW) for object storage and create an S3 -compatible bucket. ee the solution below. Solution: 1. Enable RGW: ceph mgr module enable rgw 2. Create a bucket: s3cmd mb s3://my -bucket Explanation: RADOS Gateway enables S3 and Swift -compatible object storage for Ceph, expanding storage use cases for applications. Create a storage pool with placement group auto -scaling enabled. ee the solution below. Solution: 1. Create the pool with auto -scaling: ceph osd pool create my -pool 0 ceph osd pool set my -pool pg_autoscale_mode on 2. Verify auto -scaling: ceph osd pool autoscale -status Explanation: PG auto -scaling ensures efficient data distribution across OSDs without manual intervention. Enable and configure bucket versioning for an S3 -compatible RGW bucket.
ee the solution below. Solution: 1. Enable versioning on the bucket: s3cmd versioning s3://my -bucket --enable 2. Verify versioning: s3cmd info s3://my -bucket Explanation: Bucket versioning retains multiple versions of objects, providing recovery options for accidental overwrites or deletions. Migrate data between two pools in a Ceph cluster. ee the solution below. Solution: 1. Export data from the source pool: rados export -p source -pool /backup/source -pool.dump 2. Import data into the target pool: rados import -p target -pool /backup/source -pool.dump Explanation: Data migration between pools ensures flexibility for reorganization or policy changes within the cluster. Configure the cluster to enforce read -only access for a specific client on all pools. ee the solution below. Solution: 1. Create a client with restricted permissions:
ceph auth get -or-create client.readonly mon 'allow r' osd 'allow r' 2. Verify permissions: ceph auth get client.readonly Explanation: Restricting client permissions enhances security by limiting their access to read-only operations across the cluster. Resize an existing RBD (RADOS Block Device) image. ee the solution below. Solution: 1. View the current size of the RBD: rbd info my -rbd 2. Resize the RBD: rbd resize --size 10G my -rbd 3. Verify the new size: rbd info my -rbd Explanation: Resizing RBDs allows dynamic allocation of storage, adapting to application requirements without downtime. Monitor the Ceph cluster for slow requests. ee the solution below. Solution: 1. Enable slow request logging: ceph config set mon debug_osd 20
2. Check logs for slow requests: ceph health detail Explanation: Monitoring slow requests helps identify bottlenecks or performance issues in the Ceph cluster. Set up a CephFS filesystem and mount it on a client. ee the solution below. Solution: 1. Create a CephFS filesystem: ceph fs create myfs myfs_data myfs_metadata 2. Mount the filesystem on the client: mount -t ceph <mon_ip>:/ /mnt/cephfs -o name=admin,secret=<admin_key> Explanation: CephFS provides a scalable distributed filesystem that is ideal for large -scale workloads. Recover an OSD marked as "down" or "out." ee the solution below. Solution: 1. Bring the OSD back in: ceph osd in <osd_id> 2. Restart the OSD service: systemctl restart ceph -osd@<osd_id> 3. Verify OSD status: ceph osd tree
Explanation: Recovering downed OSDs ensures that the cluster regains full operational status and redundancy. Configure Ceph to use an external key -value store for monitoring. ee the solution below. Solution: 1. Install and configure the key -value store: yum install etcd -y systemctl start etcd 2. Update the Ceph configuration: ceph config set mgr mgr/kv_backend etcd Explanation: Using an external key-value store enhances scalability and redundancy for cluster monitoring and metadata. Enable and configure Ceph cluster snapshots for data protection. ee the solution below. Solution: 1. Create a snapshot for a pool: rbd snap create my -pool/my -image@my -snapshot 2. List snapshots: rbd snap ls my -pool/my -image Explanation: Snapshots provide point -in-time recovery options for protecting against accidental data loss.
➢ TOTAL QUESTIONS: 479 Install a containerized Red Hat Ceph Storage server on a physical system using Ansible playbooks. Verify the installation. ee the solution below. Solution: 1. Install required packages: yum install ceph -ansible -y 2. Edit the Ansible inventory file (inventory.yml): [mons] node1 ansible_host=192.168.0.10 [osds] node2 ansible_host=192.168.0.11 3. Deploy the Ceph cluster: ansible -playbook -i inventory.yml site.yml 4. Verify the cluster status: ceph -s Explanation: Ansible playbooks automate the setup of a containerized Ceph Storage cluster. Verifying with ceph -s ensures the cluster is operational. Deploy a virtualized Red Hat Ceph Storage server using Podman and Ceph container images. Verify the containerized services are running. ee the solution below.