KB450430 – Adding OSD Nodes to a Ceph Cluster

Last modified: July 20, 2021
You are here:


  • This guide will detail the process of adding OSD nodes to an existing cluster running Octopus 15.2.13. The process can be completed without taking the cluster out of production.


  • An existing Ceph cluster
  • Additional OSD node(s) to add
  • The OSD node(s) have same version of Ceph installed
  • Network configured on new OSD node(s)
  • Inventory file for the new OSD node(s) configured in /usr/share/ceph-ansible/host_vars/


  • First you should check to ensure all nodes are running the same version of ceph. Use the following command to check:
rpm -qa | grep ceph
apt list --installed | grep ceph
  • Set the flags for norecover, nobackfill and norebalance.
ceph osd set norebalance 

ceph osd set nobackfill 

ceph osd set norecover
  • Make sure the cluster the flags were set:
ceph -s

  • Add the IP’s of the new OSD’s to the /etc/hosts file. For this example OSD4 is being added to the cluster.
vim /etc/hosts

  • Add passwordless ssh access to the new node(s)
ssh-copy-id root@OSD4
  • Now add the new OSD to the hosts file. Add the new OSD to the “osds” section only, and comment out the pre-existing OSD nodes.
vim /usr/share/ceph-ansible/hosts

  • Now run the following command to ensure the server can ping the new OSD(s), ensure the only OSD(s) being pinged are the ones being added to the cluster.
ansible -m ping osds
  • Once you confirm the new OSD’s can be ping you can run the core playbook to add it to the cluster.
cd /usr/share/ceph-ansible

ansible-playbook core.yml --limit osds
  • When the playbook is complete the new nodes should be added to the cluster. Run “ceph -s” to ensure it has been added to the cluster.
  • Make sure to unset the flags previously set to start the backfill process.
ceph osd unset nobackfill

ceph osd unset norecover

ceph osd unset norebalance
  • See the backfill/recovery guide to edit backfill settings to speed up/slow down backfill process.


  • If you are to run “ceph -s” you should see the new total of OSDs includes the additional nodes added to the cluster.


  • Ensure that the nodes are all running the same Ceph Version, and if need be do any minor updates to the existing Ceph Packages on the cluster so they will be in line with the new nodes added.
  • Ensure that every single disk has been added to the cluster, for example if we had 15 total OSD disks before expansion, and the new node consisted of an extra 5 drives, we should see 20 total OSD disks in the cluster.
Was this article helpful?
Dislike 0
Views: 76
Unboxing Racking Storage Drives Cable Setup Power UPS Sizing Remote Access