2. Verify the client configuration: ceph auth get client.writer Explanation: Namespace -based access control ensures that clients only interact with designated parts of the storage pool. Configure and enable placement group (PG) auto -scaling for a Ceph pool. ee the solution below. Solution: 1. Enable PG auto -scaling: ceph osd pool set my -pool pg_autoscale_mode on 2. Verify the configuration: ceph osd pool autoscale -status Explanation: PG auto -scaling dynamically adjusts the number of placement groups, ensuring optimal data distribution. Backup the Ceph configuration and restore it to a new cluster. ee the solution below. Solution: 1. Backup the configuration: ceph config dump > ceph -config.backup 2. Restore the configuration: ceph config import < ceph -config.backup Explanation:
Configuration backups safeguard against data loss and ensure quick recovery during cluster restoration. Remove an erasure -coded pool and its associated profile. ee the solution below. Solution: 1. Delete the pool: ceph osd pool delete my -ec-pool my -ec-pool --yes-i-really -really -mean -it 2. Remove the erasure code profile: ceph osd erasure -code -profile rm my -profile Explanation: Removing unused pools and profiles helps maintain a clean and organized cluster configuration. List all placement groups (PGs) and their current status. ee the solution below. Solution: 1. View PG details: ceph pg stat Explanation: Monitoring PG statuses helps identify imbalances or issues in the cluster’s data distribution. Recover data from a downed OSD using Ceph tools. ee the solution below. Solution: 1. Mark the OSD as out:
ceph osd out <osd_id> 2. Rebalance the cluster: ceph osd reweight <osd_id> 0 Explanation: Data recovery from downed OSDs maintains cluster health and prevents data loss. Monitor Ceph performance metrics using the dashboard. ee the solution below. Solution: 1. Enable the dashboard: ceph mgr module enable dashboard 2. Access the metrics: Navigate to https://<mgr_ip>:8443. Explanation: The Ceph dashboard provides a graphical view of performance metrics, simplifying cluster management. Enable and test RADOS Gateway (RGW) object storage with an S3 -compatible API. ee the solution below. Solution: 1. Enable RGW: ceph mgr module enable rgw 2. Test object storage: s3cmd put test -file s3://bucket -name Explanation:
RGW provides S3 -compatible object storage, expanding Ceph's functionality for modern applications. Expand an existing replicated pool by increasing its placement group (PG) count. ee the solution below. Solution: 1. Check the current PG count: ceph osd pool get my -pool pg_num 2. Increase the PG count: ceph osd pool set my -pool pg_num 256 3. Verify the new PG count: ceph osd pool get my -pool pg_num Explanation: Increasing the PG count improves data distribution and cluster performance but must be done gradually to avoid rebalancing overhead. Create and configure an erasure -coded pool with custom k and m values for redundancy. ee the solution below. Solution: 1. Create an erasure code profile: ceph osd erasure -code -profile set my -profile k=3 m=2 2. Create the pool using the profile: ceph osd pool create my -ec-pool 128 128 erasure my -profile Explanation: The k and m values define the data and parity fragments for redundancy, balancing storage efficiency and fault tolerance.
Configure and test restricted access for a Ceph client to a specific namespace within a pool. ee the solution below. Solution: 1. Create the restricted client: ceph auth get-or-create client.namespace -user mon 'allow r' osd 'allow rw pool=my -pool namespace=my -namespace' 2. Test access with the client: ceph -n client.namespace -user -k /etc/ceph/ceph.client.namespace -user.keyring -p my-pool -- namespace=my -namespace ls Explanation: Restricting access to specific namespaces ensures data isolation and better security for multi -tenant environments. Migrate an OSD from one CRUSH bucket to another to optimize data placement. ee the solution below. Solution: 1. View the current CRUSH map: ceph osd tree 2. Move the OSD to a new bucket: ceph osd crush move osd.<osd_id> rack=new -rack 3. Verify the changes: ceph osd tree Explanation: Moving OSDs between CRUSH buckets improves data placement and enhances fault tolerance across physical locations.
➢ TOTAL QUESTIONS: 479 Install a containerized Red Hat Ceph Storage server on a physical system using Ansible playbooks. Verify the installation. ee the solution below. Solution: 1. Install required packages: yum install ceph -ansible -y 2. Edit the Ansible inventory file (inventory.yml): [mons] node1 ansible_host=192.168.0.10 [osds] node2 ansible_host=192.168.0.11 3. Deploy the Ceph cluster: ansible -playbook -i inventory.yml site.yml 4. Verify the cluster status: ceph -s Explanation: Ansible playbooks automate the setup of a containerized Ceph Storage cluster. Verifying with ceph -s ensures the cluster is operational. Deploy a virtualized Red Hat Ceph Storage server using Podman and Ceph container images. Verify the containerized services are running. ee the solution below.