Deploy OSDs with different device technologies like lvm or physical disks using pluggable tools (lvm itself is treated like a plugin) and trying to follow a predictable, and robust way of preparing, activating, and starting OSDs.

Overview | Plugin Guide |

Command Line Subcommands There is currently support for lvm, and plain disks (with GPT partitions) that may have been deployed with ceph-disk.


Starting on Ceph version 12.2.2, ceph-disk is deprecated. Deprecation warnings will show up that will link to this page. It is strongly suggested that users start consuming ceph-volume. There are two paths for migrating:

  1. Keep OSDs deployed with ceph-disk: The simple command provides a way to take over the management while disabling ceph-disk triggers.
  2. Redeploy existing OSDs with ceph-volume: This is covered in depth on Replacing an OSD

New deployments

For new deployments, lvm is recommended, it can use any logical volume as input for data OSDs, or it can setup a minimal/naive logical volume from a device.

Existing OSDs

If the cluster has OSDs that were provisioned with ceph-disk, then ceph-volume can take over the management of these with simple. A scan is done on the data device or OSD directory, and ceph-disk is fully disabled. Encryption is fully supported.