Chapter 11. Management of Ceph OSDs on the dashboard
As a storage administrator, you can monitor and manage OSDs on the Red Hat Ceph Storage Dashboard.
Some of the capabilities of the Red Hat Ceph Storage Dashboard are:
- List OSDs, their status, statistics, information such as attributes, metadata, device health, performance counters and performance details.
- Mark OSDs down, in, out, lost, purge, reweight, scrub, deep-scrub, destroy, delete, and select profiles to adjust backfilling activity.
- List all drives associated with an OSD.
- Set and change the device class of an OSD.
- Deploy OSDs on new drives and hosts.
11.1. Prerequisites
- A running Red Hat Ceph Storage cluster
-
cluster-manager
level of access on the Red Hat Ceph Storage dashboard
11.2. Managing the OSDs on the Ceph dashboard
You can carry out the following actions on a Ceph OSD on the Red Hat Ceph Storage Dashboard:
- Create a new OSD.
- Edit the device class of the OSD.
- Mark the Flags as No Up, No Down, No In, or No Out.
- Scrub and deep-scrub the OSDs.
- Reweight the OSDs.
- Mark the OSDs Out, In, Down, or Lost.
- Purge the OSDs.
- Destroy the OSDs.
- Delete the OSDs.
Prerequisites
- A running Red Hat Ceph Storage cluster.
- Dashboard is installed.
- Hosts, Monitors and Manager Daemons are added to the storage cluster.
Procedure
- Log in to the Dashboard.
- From the Cluster drop-down menu, select OSDs.
Creating an OSD
To create the OSD, click Create.
Figure 11.1. Add device for OSDs
NoteEnsure you have an available host and a few available devices. You can check for available devices in Physical Disks under the Cluster drop-down menu.
In the Create OSDs window, from Deployment Options, select one of the below options:
- Cost/Capacity-optimized: The cluster gets deployed with all available HDDs.
- Throughput-optimized: Slower devices are used to store data and faster devices are used to store journals/WALs.
- IOPS-optmized: All the available NVMEs are used to deploy OSDs.
From the Advanced Mode, you can add primary, WAL and DB devices by clicking +Add.
- Primary devices: Primary storage devices contain all OSD data.
- WAL devices: Write-Ahead-Log devices are used for BlueStore’s internal journal and are used only if the WAL device is faster than the primary device. For example, NVMEs or SSDs.
- DB devices: DB devices are used to store BlueStore’s internal metadata and are used only if the DB device is faster than the primary device. For example, NVMEs or SSDs).
- If you want to encrypt your data for security purposes, under Features, select encryption.
- Click the Preview button and in the OSD Creation Preview dialog box, Click Create.
- In the OSD Creation Preview dialog box, Click Create.
- You get a notification that the OSD was created successfully.
- The OSD status changes from in and down to in and up.
Editing an OSD
To edit an OSD, select the row.
- From Edit drop-down menu, select Edit.
- Edit the device class.
Click Edit OSD.
Figure 11.2. Edit an OSD
- You get a notification that the OSD was updated successfully.
Marking the Flags of OSDs
To mark the flag of the OSD, select the row.
- From Edit drop-down menu, select Flags.
- Mark the Flags with No Up, No Down, No In, or No Out.
Click Update.
Figure 11.3. Marking Flags of an OSD
- You get a notification that the flags of the OSD was updated successfully.
Scrubbing the OSDs
To scrub the OSD, select the row.
- From Edit drop-down menu, select Scrub.
In the OSDs Scrub dialog box, click Update.
Figure 11.4. Scrubbing an OSD
- You get a notification that the scrubbing of the OSD was initiated successfully.
Deep-scrubbing the OSDs
To deep-scrub the OSD, select the row.
- From Edit drop-down menu, select Deep scrub.
In the OSDs Deep Scrub dialog box, click Update.
Figure 11.5. Deep-scrubbing an OSD
- You get a notification that the deep scrubbing of the OSD was initiated successfully.
Reweighting the OSDs
To reweight the OSD, select the row.
- From Edit drop-down menu, select Reweight.
- In the Reweight OSD dialog box, enter a value between zero and one.
Click Reweight.
Figure 11.6. Reweighting an OSD
Marking OSDs Out
To mark the OSD out, select the row.
- From Edit drop-down menu, select Mark Out.
In the Mark OSD out dialog box, click Mark Out.
Figure 11.7. Marking OSDs out
- The status of the OSD will change to out.
Marking OSDs In
To mark the OSD in, select the OSD row that is in out status.
- From Edit drop-down menu, select Mark In.
In the Mark OSD in dialog box, click Mark In.
Figure 11.8. Marking OSDs in
- The status of the OSD will change to in.
Marking OSDs Down
To mark the OSD down, select the row.
- From Edit drop-down menu, select Mark Down.
In the Mark OSD down dialog box, click Mark Down.
Figure 11.9. Marking OSDs down
- The status of the OSD will change to down.
Marking OSDs Lost
To mark the OSD lost, select the OSD in out and down status.
- From Edit drop-down menu, select Mark Lost.
In the Mark OSD Lost dialog box, check Yes, I am sure option, and click Mark Lost.
Figure 11.10. Marking OSDs Lost
Purging OSDs
To purge the OSD, select the OSD in down status.
- From Edit drop-down menu, select Purge.
In the Purge OSDs dialog box, check Yes, I am sure option, and click Purge OSD.
Figure 11.11. Purging OSDs
- All the flags are reset and the OSD is back in in and up status.
Destroying OSDs
To destroy the OSD, select the OSD in down status.
- From Edit drop-down menu, select Destroy.
In the Destroy OSDs dialog box, check Yes, I am sure option, and click Destroy OSD.
Figure 11.12. Destroying OSDs
- The status of the OSD changes to destroyed.
Deleting OSDs
To delete the OSD, select the OSD in down status.
- From Edit drop-down menu, select Delete.
In the Destroy OSDs dialog box, check Yes, I am sure option, and click Delete OSD.
NoteYou can preserve the OSD_ID when you have to to replace the failed OSD.
Figure 11.13. Deleting OSDs
11.3. Replacing the failed OSDs on the Ceph dashboard
You can replace the failed OSDs in a Red Hat Ceph Storage cluster with the cluster-manager
level of access on the dashboard. One of the highlights of this feature on the dashboard is that the OSD IDs can be preserved while replacing the failed OSDs.
Prerequisites
- A running Red Hat Ceph Storage cluster.
-
At least
cluster-manager
level of access to the Ceph Dashboard. -
At least one of the OSDs is
down
Procedure
On the dashboard, you can identify the failed OSDs in the following ways:
- Dashboard AlertManager pop-up notifications.
- Dashboard landing page showing HEALTH_WARN status.
- Dashboard landing page showing failed OSDs.
Dashboard OSD page showing failed OSDs.
In this example, you can see that one of the OSDs is down on the landing page of the dashboard.
Apart from this, on the physical drive, you can view the LED lights blinking if one of the OSDs is down.
- Click OSDs.
Select the
out
anddown
OSD:- From the Edit drop-down menu, select Flags and select No Up and click Update.
- From the Edit drop-down menu, select Delete.
- In the Delete OSD dialog box, select the Preserve OSD ID(s) for replacement and Yes, I am sure check boxes.
- Click Delete OSD.
-
Wait till the status of the OSD changes to
out
anddestroyed
status.
Optional: If you want to change the No Up Flag for the entire cluster, in the Cluster-wide configuration drop-down menu, select Flags.
- In Cluster-wide OSDs Flags dialog box, select No Up and click Update.
Optional: If the OSDs are down due to a hard disk failure, replace the physical drive:
- If the drive is hot-swappable, replace the failed drive with a new one.
- If the drive is not hot-swappable and the host contains multiple OSDs, you might have to shut down the whole host and replace the physical drive. Consider preventing the cluster from backfilling. See the Stopping and Starting Rebalancing chapter in the Red Hat Ceph Storage Troubleshooting Guide for details.
-
When the drive appears under the
/dev/
directory, make a note of the drive path. - If you want to add the OSD manually, find the OSD drive and format the disk.
If the new disk has data, zap the disk:
Syntax
ceph orch device zap HOST_NAME PATH --force
Example
ceph orch device zap ceph-adm2 /dev/sdc --force
- From the Create drop-down menu, select Create.
In the Create OSDs window, click +Add for Primary devices.
In the Primary devices dialog box, from the Hostname drop-down list, select any one filter. From Any drop-down list, select the respective option.
NoteYou have to select the Hostname first and then at least one filter to add the devices.
For example, from Hostname list, select
Type
and from Any list selecthdd
. Select Vendor and from Any list, selectATA
- Click Add.
- In the Create OSDs window, click the Preview button.
- In the OSD Creation Preview dialog box, Click Create.
-
You will get a notification that the OSD is created. The OSD will be in
out
anddown
status.
Select the newly created OSD that has out and down status.
- In the Edit drop-down menu, select Mark-in.
- In the Mark OSD in window, select Mark in.
- In the Edit drop-down menu, select Flags.
- Uncheck No Up and click Update.
Optional: If you have changed the No Up Flag before for cluster-wide configuration, in the Cluster-wide configuration menu, select Flags.
- In Cluster-wide OSDs Flags dialog box, uncheck No Up and click Update.
Verification
Verify that the OSD that was destroyed is created on the device and the OSD ID is preserved.
Additional Resources
- For more information on Down OSDs, see the Down OSDs section in the Red Hat Ceph Storage Troubleshooting Guide.
- For additional assistance see the Red Hat Support for service section in the Red Hat Ceph Storage Troubleshooting Guide.
- For more information on system roles, see the User roles and permissions on the Ceph dashboard section in the Red Hat Ceph Storage Dashboard Guide.