A Ceph cluster can be made up of 10 to several thousand physical disks that provide storage capacity to the cluster. As the number of physical disks increases for your Ceph cluster, the frequency of disk failures also increases. Hence, replacing a failed disk drive might become a repetitive task for a Ceph storage administrator. In this recipe, we will learn about the disk replacement process for a Ceph cluster.
Let's verify cluster health; since this cluster does not have any failed disk status, it would be
HEALTH_OK
:# ceph status
Since we are demonstrating this exercise on virtual machines, we need to forcefully fail a disk by bringing
ceph-node1
down, detaching a disk, and powering up the VM. Execute the following commands from your HOST machine:# VBoxManage controlvm ceph-node1 poweroff # VBoxManage storageattach ceph-node1 --storagectl "SATA" --port 1 --device 0 --type hdd --medium none # VBoxManage startvm ceph-node1
The following...