Chapter 11. Managing Ceph OSDs on the dashboard


As a storage administrator, you can monitor and manage OSDs on the Red Hat Ceph Storage Dashboard.

Some of the capabilities of the Red Hat Ceph Storage Dashboard are:

  • List OSDs, their status, statistics, information such as attributes, metadata, device health, performance counters and performance details.
  • Mark OSDs down, in, out, lost, purge, reweight, scrub, deep-scrub, destroy, delete, and select profiles to adjust backfilling activity.
  • List all drives associated with an OSD.
  • Set and change the device class of an OSD.
  • Deploy OSDs on new drives and hosts.

Prerequisites

  • A running Red Hat Ceph Storage cluster
  • cluster-manager level of access on the Red Hat Ceph Storage dashboard

11.1. Managing the OSDs on the Ceph dashboard

You can carry out the following actions on a Ceph OSD on the Red Hat Ceph Storage Dashboard:

  • Create a new OSD.
  • Edit the device class of the OSD.
  • Mark the Flags as No Up, No Down, No In, or No Out.
  • Scrub and deep-scrub the OSDs.
  • Reweight the OSDs.
  • Mark the OSDs Out, In, Down, or Lost.
  • Purge the OSDs.
  • Destroy the OSDs.
  • Delete the OSDs.

Prerequisites

  • A running Red Hat Ceph Storage cluster.
  • Dashboard is installed.
  • Hosts, Monitors, and Manager Daemons are added to the storage cluster.

Procedure

From the dashboard navigation, go to Cluster→OSDs.

Creating an OSD

  1. To create the OSD, from the OSDs List table, click Create.

    Figure 11.1. Add device for OSDs

    Add device for OSDs
    Note

    Ensure you have an available host and a few available devices. Check for available devices in Cluster→Physical Disks and filter for Available.

    1. In the Create OSDs form, in the Deployment Options section, select one of the following options:

      • Cost/Capacity-optimized: The cluster gets deployed with all available HDDs.
      • Throughput-optimized: Slower devices are used to store data and faster devices are used to store journals/WALs.
      • IOPS-optmized: All the available NVMe devices are used to deploy OSDs.
    2. In the Advanced Mode section, add primary, WAL, and DB devices by clicking Add.

      • Primary devices: Primary storage devices contain all OSD data.
      • WAL devices: Write-Ahead-Log devices are used for BlueStore’s internal journal and are used only if the WAL device is faster than the primary device. For example, NVMe or SSD devices.
      • DB devices: DB devices are used to store BlueStore’s internal metadata and are used only if the DB device is faster than the primary device. For example, NVMe or SSD devices.
    3. To encrypt your data, for security purposes, from the Features section of the form, select Encryption.
    4. Click Preview.
    5. In the OSD Creation Preview dialog review the OSD and click Create.

      A notification displays that the OSD was created successfully and the OSD status changes from in and down to in and up.

Editing an OSD

  1. To edit an OSD, select the row and click Edit.

    1. From the Edit OSD form, edit the device class.
    2. Click Edit OSD.

      Figure 11.2. Edit an OSD

      Edit an OSD

      A notification displays that the OSD was updated successfully.

Marking the OSD flags

  1. To mark the flag of the OSD, select the row and click Flags from the action drop-down.
  2. In the Individual OSD Flags form, select the OSD flags needed.
  3. Click Update.

    Figure 11.3. Marking OSD flags

    Marking Flags of an OSD

    A notification displays that the OSD flags updated successfully.

Scrubbing an OSD

  1. To scrub an OSD, select the row and click Scrub from the action drop-down.
  2. In the OSDs Scrub notification, click Update.

    Figure 11.4. Scrubbing an OSD

    Scrubbing an OSD

    A notification displays that the scrubbing of the OSD was initiated successfully.

Deep-scrubbing the OSDs

  1. To deep-scrub the OSD, select the row and click Deep Scrub from the action drop-down.
  2. In the OSDs Deep Scrub notification, click Update.

    Figure 11.5. Deep-scrubbing an OSD

    Deep-scrubbing an OSD

    A notification displays that the deep scrubbing of the OSD was initiated successfully.

Reweighting the OSDs

  1. To reweight the OSD, select the row and click Reweight from the action drop-down.
  2. In the Reweight OSD form enter a value between 0 and 1.
  3. Click Reweight.

    Figure 11.6. Reweighting an OSD

    Reweighting an OSD

Marking OSDs out

  1. To mark an OSD as out, select the row and click Mark Out from the action drop-down.
  2. In the Mark OSD out notification, click Mark Out.

    Figure 11.7. Marking OSDs out

    Marking OSDs out

    The OSD status changes to out.

Marking OSDs in

  1. To mark an OSD as in, select the OSD row that is in out status and click Mark In from the action drop-down.
  2. In the Mark OSD in notification, click Mark In.

    Figure 11.8. Marking OSDs in

    Marking OSDs in

    The OSD status changes to in.

Marking OSDs down

  1. To mark an OSD down, select the row and click Mark Down from the action drop-down.
  2. In the Mark OSD down notification, click Mark Down.

    Figure 11.9. Marking OSDs down

    Marking OSDs down

    The OSD status changes to down.

Marking OSDs lost

  1. To mark an OSD lost, select the OSD in out and down status and click Mark Lost from the action drop-down.
  2. In the Mark OSD Lost notification, select Yes, I am sure and click Mark Lost.

    Figure 11.10. Marking OSDs lost

    Marking OSDs lost

Purging OSDs

  1. To purge an OSD, select the OSD in down status and click Purge from the action drop-down.
  2. In the Purge OSDs notification, select Yes, I am sure and click Purge OSD.

    Figure 11.11. Purging OSDs

    Purging OSDs

    All the flags are reset and the OSD is back in in and up status.

Destroying OSDs

  1. To destroy an OSD, select the OSD in down status and click Destroy from the action drop-down.
  2. In the Destroy OSDs notification, select Yes, I am sure and click Destroy OSD.

    Figure 11.12. Destroying OSDs

    Destroying OSDs

    The OSD status changes to destroyed.

Deleting OSDs

  1. To delete an OSD, select the OSD and click Delete from the action drop-down.
  2. In the Delete OSDs notification, select Yes, I am sure and click Delete OSD.

    Note

    You can preserve the OSD_ID when you have to to replace the failed OSD.

    Figure 11.13. Deleting OSDs

    Deleting OSDs

11.2. Replacing the failed OSDs on the Ceph dashboard

You can replace the failed OSDs in a Red Hat Ceph Storage cluster with the cluster-manager level of access on the dashboard. One of the highlights of this feature on the dashboard is that the OSD IDs can be preserved while replacing the failed OSDs.

Prerequisites

  • A running Red Hat Ceph Storage cluster.
  • At least cluster-manager level of access to the Ceph Dashboard.
  • At least one of the OSDs is down

Procedure

  1. On the dashboard, you can identify the failed OSDs in the following ways:

    • Dashboard AlertManager pop-up notifications.
    • Dashboard landing page showing HEALTH_WARN status.
    • Dashboard landing page showing failed OSDs.
    • Dashboard OSD list showing failed OSDs.

      In the following example, you can see that one of the OSDs is down and one is out on the landing page of the dashboard.

      Figure 11.14. OSD status on the Ceph Dashboard landing page

      Health status of OSDs

      You can also view the LED blinking lights on the physical drive if one of the OSDs is down.

  2. From Cluster→OSDs, on the OSDs List table, select the out and down OSD.

    1. Click Flags from the action drop-down, select No Up in the Individual OSD Flags form, and click Update.
    2. Click Delete from the action drop-down. In the Delete OSD notification, select Preserve OSD ID(s) for replacement and Yes, I am sure and click Delete OSD.
    3. Wait until the status of the OSD changes to out and destroyed.
  3. Optional: To change the No Up Flag for the entire cluster, from the Cluster-wide configuration menu, select Flags.

    1. In Cluster-wide OSDs Flags form, select No Up and click Update.
  4. Optional: If the OSDs are down due to a hard disk failure, replace the physical drive:

    • If the drive is hot-swappable, replace the failed drive with a new one.
    • If the drive is not hot-swappable and the host contains multiple OSDs, you might have to shut down the whole host and replace the physical drive. Consider preventing the cluster from backfilling. See the Stopping and Starting Rebalancing chapter in the Red Hat Ceph Storage Troubleshooting Guide for details.
    • When the drive appears under the /dev/ directory, make a note of the drive path.
    • If you want to add the OSD manually, find the OSD drive and format the disk.
    • If the new disk has data, zap the disk:

      Syntax

      ceph orch device zap HOST_NAME PATH --force

      Example

      ceph orch device zap ceph-adm2 /dev/sdc --force

  5. From the Ceph Dashboard OSDs List, click Create.
  6. In the Create OSDs form Advanced Mode section, add a primary device.

    1. In the Primary devices dialog, select a Hostname filter.
    2. Select a device type from the list.

      Note

      You have to select the Hostname first and then at least one filter to add the devices.

      For example, from Hostname list, select Type and then hdd.

    3. Select Vendor and from device list, select ATA.

      Figure 11.15. Using the Primary devices filter

      Add device for OSDs
    4. Click Add.
    5. In the Create OSDs form, click Preview.
    6. In the OSD Creation Preview dialog, click Create.

      A notification displays that the OSD is created successfully and the OSD changes to be in the out and down status.

  7. Select the newly created OSD that has out and down status.

    1. Click Mark In from the action drop-down.
    2. In the Mark OSD in notification, click Mark In.

      The OSD status changes to in.

    3. Click Flags from the action drop-down.
    4. Clear the No Up selection and click Update.
  8. Optional: If you have changed the No Up flag before for cluster-wide configuration, in the Cluster-wide configuration menu, select Flags.

    1. In Cluster-wide OSDs Flags form, clear the No Up selection and click Update.

Verification

  1. Verify that the OSD that was destroyed is created on the device and the OSD ID is preserved.

    OSD is created

Additional Resources

  • For more information on Down OSDs, see the Down OSDs section in the Red Hat Ceph Storage Troubleshooting Guide.
  • For additional assistance see the Red Hat Support for service section in the Red Hat Ceph Storage Troubleshooting Guide.
  • For more information on system roles, see the Managing roles on the Ceph dashboard section in the Red Hat Ceph Storage Dashboard Guide.
Red Hat logoGithubRedditYoutubeTwitter

Learn

Try, buy, & sell

Communities

About Red Hat Documentation

We help Red Hat users innovate and achieve their goals with our products and services with content they can trust.

Making open source more inclusive

Red Hat is committed to replacing problematic language in our code, documentation, and web properties. For more details, see the Red Hat Blog.

About Red Hat

We deliver hardened solutions that make it easier for enterprises to work across platforms and environments, from the core datacenter to the network edge.

© 2024 Red Hat, Inc.