Deploy OSDs with different device technologies like lvm or physical disks using pluggable tools (lvm itself is treated like a plugin) and trying to follow a predictable, and robust way of preparing, activating, and starting OSDs.

Overview | Plugin Guide |

Command Line Subcommands

There is currently support for lvm, and plain disks (with GPT partitions) that may have been deployed with ceph-disk.

zfs support is available for running a FreeBSD cluster.

Node inventory

The inventory subcommand provides information and metadata about a nodes physical disk inventory.


Starting on Ceph version 13.0.0, ceph-disk is deprecated. Deprecation warnings will show up that will link to this page. It is strongly suggested that users start consuming ceph-volume. There are two paths for migrating:

  1. Keep OSDs deployed with ceph-disk: The simple command provides a way to take over the management while disabling ceph-disk triggers.

  2. Redeploy existing OSDs with ceph-volume: This is covered in depth on Replacing an OSD

For details on why ceph-disk was removed please see the Why was ceph-disk replaced? section.

New deployments

For new deployments, lvm is recommended, it can use any logical volume as input for data OSDs, or it can setup a minimal/naive logical volume from a device.

Existing OSDs

If the cluster has OSDs that were provisioned with ceph-disk, then ceph-volume can take over the management of these with simple. A scan is done on the data device or OSD directory, and ceph-disk is fully disabled. Encryption is fully supported.