➢ TOTAL QUESTIONS: 479 Install a containerized Red Hat Ceph Storage server on a physical system using Ansible playbooks. Verify the installation. ee the solution below. Solution: 1. Install required packages: yum install ceph -ansible -y 2. Edit the Ansible inventory file (inventory.yml): [mons] node1 ansible_host=192.168.0.10 [osds] node2 ansible_host=192.168.0.11 3. Deploy the Ceph cluster: ansible -playbook -i inventory.yml site.yml 4. Verify the cluster status: ceph -s Explanation: Ansible playbooks automate the setup of a containerized Ceph Storage cluster. Verifying with ceph -s ensures the cluster is operational. Deploy a virtualized Red Hat Ceph Storage server using Podman and Ceph container images. Verify the containerized services are running. ee the solution below.
ee the solution below. Solution: 1. Enable snapshot scheduling: ceph mgr module enable rbd_support 2. Configure the snapshot schedule: rbd snap -schedule add my -pool/my -rbd --interval 6h 3. Verify the snapshot schedule: rbd snap -schedule status my -pool/my -rbd Explanation: Automatic snapshot scheduling ensures consistent backups, reducing the risk of data loss in production environments. Configure RBD mirroring in pool mode, set up a new peer cluster, and ensure all images are synchronized. ee the solution below. Solution: 1. Enable mirroring on the pool: rbd mirror pool enable my -pool pool 2. Add a new peer cluster: rbd mirror pool peer add my -pool client.admin@remote -cluster 3. Verify synchronization status: rbd mirror pool status my -pool Explanation: Pool mode mirroring replicates all images in the pool to the peer cluster, ensuring disaster recovery readiness.
Create an RBD image with a specific feature set and verify the enabled features. ee the solution below. Solution: 1. Create an RBD image with specified features: rbd create my -rbd --size 10240 --pool my -pool --image -feature exclusive -lock,fast -diff,deep -flatten 2. Verify the features: rbd info my -pool/my -rbd Explanation: Specifying features like exclusive -lock, fast-diff, and deep -flatten ensures the RBD image is optimized for performance and manageability. Create an RBD snapshot, export it as a differential file, and apply it to an existing base image in a different pool. ee the solution below. Solution: 1. Create a snapshot: rbd snap create my -pool/my -rbd@my -snapshot 2. Export the snapshot as a differential file: rbd export -diff my -pool/my -rbd@my -snapshot /backup/my -snapshot.diff 3. Apply the differential file to the base image in another pool: rbd import -diff /backup/my -snapshot.diff new -pool/my -base Explanation: Differential exports save only changes, enabling efficient synchronization and recovery of images in different pools.
Enable and test pool -level quotas for both maximum data usage and the number of objects. ee the solution below. Solution: 1. Set a maximum data usage quota: ceph osd pool set -quota my -pool max_bytes 10737418240 2. Set a maximum object quota: ceph osd pool set -quota my -pool max_objects 1000 3. Verify the quotas: ceph osd pool get -quota my -pool Explanation: Quotas prevent excessive resource consumption, ensuring fair usage and better cluster management. Configure two-way RBD mirroring in image mode for selected images in a pool and monitor the synchronization status. ee the solution below. Solution: 1. Enable mirroring for selected images: rbd mirror image enable my -pool/my -image 2. Add a peer for the image: rbd mirror image peer add my -pool/my -image client.admin@remote -cluster 3. Verify synchronization status: rbd mirror image status my -pool/my -image Explanation:
Two-way mirroring ensures selected images are synchronized between clusters, supporting high availability and disaster recovery. Automate RBD snapshot creation and deletion using a custom schedule and verify the operations. ee the solution below. Solution: 1. Enable snapshot scheduling: ceph mgr module enable rbd_support 2. Set up the snapshot schedule: rbd snap -schedule add my -pool/my -image --interval 24h --retention 7 3. Verify the schedule and retention settings: rbd snap -schedule status my -pool/my -image Explanation: Automating snapshot creation and retention reduces manual efforts and ensures consistent backups. Resize an RBD image online and expand its filesystem without unmounting it. ee the solution below. Solution: 1. Resize the RBD image: rbd resize --size 30720 my -pool/my -image 2. Expand the filesystem: xfs_growfs /mnt/data 3. Verify the new size: df -h | grep /mnt/data
3. Apply the new CRUSH map: crushtool -c crush.txt -o crush.map ceph osd setcrushmap -i crush.map Explanation: CRUSH maps define how data is distributed across OSDs. Customizing these maps allows fine-grained control over data placement. Set up Ceph storage with authentication enabled and verify that clients require keys to access the cluster. ee the solution below. Solution: 1. Enable authentication in the Ceph configuration file: ceph config set mon auth_allow_insecure_global_id_reclaim false 2. Generate a client key: ceph auth get -or-create client.admin 3. Verify client access using the key: ceph -s --keyring /etc/ceph/ceph.client.admin.keyring Explanation: Enabling authentication adds a layer of security, ensuring only authorized clients can interact with the cluster. Deploy Ceph storage and configure BlueStore as the OSD backend for optimal performance. ee the solution below. Solution: 1. Update the playbook with BlueStore settings: osd_objectstore: bluestore