The ODF CLI tool provides configuration and troubleshooting commands for OpenShift Data Foundation.
odf set
:recovery-profile <profile>
: Set the recovery profile to favor new IO, recovery, or balanced mode with optionshigh_client_ops
,high_recovery_ops
, orbalanced
. The default isbalanced
.full <ratio>
: Update the ceph osd full ratio after which Ceph automatically prevents any I/O operations on OSDs.nearfull <ratio>
: Update the ceph osd nearfull ratio in case Ceph returns the nearfull osds message when the cluster reaches the capacity specified.backfillfull <ratio>
: Update the ceph osd backfillfull ratio in case ceph will deny backfilling to the OSD that reached the capacity specified.ceph
log-level <daemon> <subsystem> <log-level>
: Set the log level for Ceph daemons like OSD, mon, mds etc. More information about the ceph subsystems can be found here
odf get
:recovery-profile
: Get the recovery profile value.health
: Check health of the cluster and common configuration issues.dr-health [ceph status args]
: Print the ceph status of a peer cluster in a mirroring-enabled environment thereby validating connectivity between ceph clusters. Ceph status args can be optionally passed, such as to change the log level: --debug-ms 1.dr-prereq <PeerManagedClusterName>
: Print the status of pre-requisites for Disaster Recovery between peer clusters.mon-endpoints
: Print mon endpoints.
odf purge-osd <ID>
: Permanently remove an OSD from the cluster.odf maintenance
: Perform maintenance operations on mons or OSDs. The mon or OSD deployment will be scaled down and replaced temporarily by a maintenance deployment.start <deployment-name>
[--alternate-image <alternate-image>]
: (optional) Start a maintenance deployment with an optional alternative ceph container image
stop <deployment-name>
: Stop the maintenance deployment and restore the mon or OSD deployment
odf subvolume
:ls
: Display all the subvolumesdelete <subvolume> <filesystem> <subvolumegroup>
: Deletes the stale subvolumes
odf operator
:rook
:set
: Set the property in the rook-ceph-operator-config configmap.restart
: Restart the Rook-Ceph operator
odf restore
:mon-quorum
: Restore the mon quorum based on a single healthy mon since quorum was lost with the other monsdeleted
: Restore the ceph resources which are stuck in deleting state due to underlying resources being present in the cluster
odf ceph <args>
: Run a Ceph CLI command. Supports any arguments the ceph command supports. See Ceph docs for more.odf rados <args>
: Run a Rados CLI command. Supports any arguments the rados command supports. See Rados docs for more.odf radosgw-admin <args>
: Run an RGW CLI command. Supports any arguments the radosgw-admin command supports. See the radosgw-admin docs for more.odf rbd <args>
: Call a 'rbd' CLI command with arbitrary argsodf help
: Display help text
Visit docs below for complete details about each command and their flags uses.
These are the arguments that apply to all commands:
-
-h|--help
: this will print brief command help text.odf -h
-
-n|--namespace='openshift-storage'
: the Openshift namespace in which the StorageCluster resides. (optional, default: openshift-storage)odf -n test-cluster [commands]
-
-o|--operator-namespace
: the Openshift namespace in which the rook operator resides, when the arg-n
is passed but-o
is not then-o
will equal to the-n
. (default: openshift-storage)odf --operator-namespace test-operator -n test-cluster [commands]
-
--context
: the name of the Openshift context to be used (optional).odf --context=$(oc config current-context) [commands]
- Go >= 1.22.0
- ODF storage cluster should be installed.
-
Clone the repository
git clone https://github.com/red-hat-storage/odf-cli.git
-
Change the directory and build the binary
cd odf-cli/ && make build
-
Use the binary present in the
/bin/
directory to run the commands./bin/odf -h