Chapter 1. Configuring a Hyperconverged Infrastructure environment
This section describes how to deploy a Hyperconverged Infrastructure (HCI) environment. A HCI environment contains data plane nodes that host both Ceph Storage and the Compute service.
Create an HCI environment, by completing the following high-level tasks:
- Configuring the data plane node networking.
- Installing Red Hat Ceph Storage on the data plane nodes.
- Configuring Red Hat OpenStack Services on OpenShift (RHOSO) to use the Red Hat Ceph Storage cluster.
Red Hat OpenStack Services on OpenShift (RHOSO) supports external deployments of Red Hat Ceph Storage 7 and 8. Configuration examples that reference Red Hat Ceph Storage use Release 7 information. If you are using Red Hat Ceph Storage 8, adjust the configuration examples accordingly.
1.1. Data plane node services list Copy linkLink copied to clipboard!
Create an OpenStackDataPlaneNodeSet CR to configure data plane nodes. The openstack-operator reconciles the OpenStackDataPlaneNodeSet CR when an OpenStackDataPlaneDeployment CR is created.
These CRs have a service list similar to the following example:
This CR representation is an example and may not list all of the services in your environment. For a default list of services in your environment, use the following command:
oc get -n openstack crd/openstackdataplanenodesets.dataplane.openstack.org -o yaml |yq -r '.spec.versions.[].schema.openAPIV3Schema.properties.spec.properties.services.default
oc get -n openstack crd/openstackdataplanenodesets.dataplane.openstack.org -o yaml |yq -r '.spec.versions.[].schema.openAPIV3Schema.properties.spec.properties.services.default
For more information, see Data plane services.
Only the services in the services list are configured.
Red Hat Ceph Storage must be deployed on the data plane node after the Storage network and NTP are configured but before the Compute service is configured. This means you must edit the services list and make other changes to the CR. Throughout this section, you edit the services list to complete the configuration of the HCI environment.
1.2. Configuring the data plane node networks Copy linkLink copied to clipboard!
You must configure the data plane node networks to accommodate the Red Hat Ceph Storage networking requirements.
Prerequisites
- Control plane deployment is complete but has not yet been modified to use Ceph Storage.
- The data plane nodes have been provisioned with an operating system.
- The data plane nodes are accessible through an SSH key that Ansible can use.
- The data plane nodes have disks available to be used as Ceph OSDs.
- There are a minimum of three available data plane nodes. Ceph Storage clusters must have a minimum of three nodes to ensure redundancy.
Procedure
Create an
OpenStackDataPlaneNodeSetCRD file to represent the data plane nodes.NoteDo not create the CR in Red Hat OpenShift yet.
Add the
ceph-hci-preservice to the list of services before theconfigure-osservice, and remove all other service listings afterreboot-os.The following is an example of the edited list:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow NoteNote the services that you remove from the list. You add them back to the list later.
-
(Optional) The
ceph-hci-preservice prepares EDPM nodes to host Red Hat Ceph Storage services after network configuration using theedpm_ceph_hci_pre edpm-ansiblerole. By default, theedpm_ceph_hci_pre_enabled_servicesparameter of this role only contains RBD, RGW, and NFS services. If other services, such as the Dashboard, are deployed with HCI nodes; they must be added to theedpm_ceph_hci_pre_enabled_servicesparameter list. For more information about this role, see edpm_ceph_hci_pre role. Configure the Red Hat Ceph Storage
cluster_networkfor storage management traffic between OSDs. Modify the CR to setedpm-ansiblevariables so that theedpm_network_configrole configures a storage management network which Ceph uses as thecluster_network.The following example has 3 nodes. It assumes the storage management network range is
172.20.0.0/24and that it is onVLAN23. The bolded lines are additions for thecluster_network:Copy to Clipboard Copied! Toggle word wrap Toggle overflow NoteIt is not necessary to add the storage management network to the networkAttachments key.
Apply the CR:
oc apply -f <dataplane_cr_file>
$ oc apply -f <dataplane_cr_file>Copy to Clipboard Copied! Toggle word wrap Toggle overflow Replace
<dataplane_cr_file>with the name of your file.NoteAnsible does not configure or validate the networks until the
OpenStackDataPlaneDeploymentCRD is created.
-
Create an
OpenStackDataPlaneDeploymentCRD, as described in Creating the data plane in the Deploying Red Hat OpenStack Services on OpenShift guide, which has theOpenStackDataPlaneNodeSetCRD file defined above to have Ansible configure the services on the data plane nodes. To confirm the network is configured, complete the following steps:
- SSH into a data plane node.
-
Use the
ip acommand to display configured networks. - Confirm the storage networks are in the list of configured networks.
1.2.1. Red Hat Ceph Storage MTU settings Copy linkLink copied to clipboard!
The example in this procedure changes the MTU of the storage and storage_mgmt networks from 1500 to 9000. An MTU of 9000 is known as a jumbo frame. Even though it is not mandatory to increase the MTU, jumbo frames are used for improved storage performance. If jumbo frames are used, all network switch ports in the data path must be configured to support jumbo frames. MTU changes must also be made for services using the Storage network running on OpenShift.
To change the MTU for the OpenShift services connecting to the data plane nodes, update the Node Network Configuration Policy (NNCP) for the base interface and the VLAN interface. It is not necessary to update the Network Attachment Definition (NAD) if the main NAD interface already has the desired MTU. If the MTU of the underlying interface is set to 9000, and it is not specified for the VLAN interface on top of it, then it will default to the value from the underlying interface.
If the MTU values are not consistent, issues can occur on the application layer that can cause the Red Hat Ceph Storage cluster to not reach quorum or not support authentication using the CephX protocol. If the MTU is changed and you observe these types of problems, verify all hosts that use the network using jumbo frames can communicate at the chosen MTU value with the ping command, for example:
ping -M do -s 8972 172.20.0.100
$ ping -M do -s 8972 172.20.0.100
1.3. Configuring and deploying Red Hat Ceph Storage on data plane nodes Copy linkLink copied to clipboard!
Use the cephadm utility to configure and deploy Red Hat Ceph Storage for an HCI environment.
1.3.1. The cephadm utility Copy linkLink copied to clipboard!
Use the cephadm utility to configure and deploy Red Hat Ceph Storage on the data plane nodes. The cephadm package must be deployed on at least one data plane node before proceeding; edpm-ansible does not deploy Red Hat Ceph Storage.
For additional information and procedures for deploying Red Hat Ceph Storage, see Red Hat Ceph Storage installation in the Red Hat Ceph Storage Installation Guide.
1.3.2. Configuring and deploying Red Hat Ceph Storage Copy linkLink copied to clipboard!
Configure and deploy Red Hat Ceph Storage by editing the configuration file and using the cephadm utility.
In HCI deployments where Red Hat Ceph Storage is co-located with Compute (nova) nodes, IP addresses on the storage and storage_mgmt networks must remain fixed after initial deployment. For more information, see Fixed IP addressing for HCI deployments.
Procedure
- Edit the Red Hat Ceph Storage configuration file.
Add the
StorageandStorage Managementnetwork ranges. Red Hat Ceph Storage uses theStoragenetwork as the Red Hat Ceph Storagepublic_networkand theStorage Managementnetwork as thecluster_network.The following example is for a configuration file entry where the
Storagenetwork range is172.18.0.0/24and theStorage Managementnetwork range is172.20.0.0/24:[global] public_network = 172.18.0.0/24 cluster_network = 172.20.0.0/24
[global] public_network = 172.18.0.0/24 cluster_network = 172.20.0.0/24Copy to Clipboard Copied! Toggle word wrap Toggle overflow Add co-location boundaries between the Compute service and Ceph OSD services. Boundaries should be set between co-located Compute service and Ceph OSD services to reduce CPU and memory contention.
The following is an example for a Ceph configuration file entry with these boundaries set:
[osd] osd_memory_target_autotune = true osd_numa_auto_affinity = true [mgr] mgr/cephadm/autotune_memory_target_ratio = 0.2
[osd] osd_memory_target_autotune = true osd_numa_auto_affinity = true [mgr] mgr/cephadm/autotune_memory_target_ratio = 0.2Copy to Clipboard Copied! Toggle word wrap Toggle overflow In this example, the
osd_memory_target_autotuneparameter is set totrueso that the OSD daemons adjust memory consumption based on theosd_memory_targetoption. Theautotune_memory_target_ratiodefaults to 0.7. This means 70 percent of the total RAM in the system is the starting point from which any memory consumed by non-autotuned Ceph daemons is subtracted. The remaining memory is divided between the OSDs; assuming all OSDs haveosd_memory_target_autotuneset to true. For HCI deployments, you can setmgr/cephadm/autotune_memory_target_ratioto 0.2 so that more memory is available for the Compute service.For additional information about service co-location, see Co-locating services in a HCI environment for NUMA nodes.
NoteIf these values need to be adjusted after the deployment, use the
ceph config set osd <key> <value>command.Deploy Ceph Storage with the edited configuration file on a data plane node:
$ cephadm bootstrap --config <config_file> --mon-ip <data_plane_node_ip> --skip-monitoring-stack-
Replace
<config_file>with the name of your Ceph configuration file. Replace
<data_plane_node_ip>with theStoragenetwork IP address of the data plane node on which Red Hat Ceph Storage will be installed.NoteThe
--skip-monitoring-stackoption is used in thecephadm bootstrapcommand to skip the deployment of monitoring services. This ensures the Red Hat Ceph Storage deployment completes successfully if monitoring services have been previously deployed as part of any other preceding process.If monitoring services have not been deployed, see the Red Hat Ceph Storage documentation for information and procedures on enabling monitoring services.
-
Replace
- After the Red Hat Ceph Storage cluster is bootstrapped on the first EDPM node, see Red Hat Ceph Storage installation in the Red Hat Ceph Storage Installation Guide to add the other EDPM nodes to the Ceph cluster.
1.3.2.1. Fixed IP addressing for HCI deployments Copy linkLink copied to clipboard!
In HCI deployments where Red Hat Ceph Storage is co-located with Compute (nova) nodes, IP addresses on the storage and storage_mgmt networks must remain fixed after initial deployment.
Workloads use the IP addresses on the storage network (Ceph public network) to connect to Ceph monitors (MONs). Changing these addresses breaks client connectivity to the storage cluster. Ceph Object Storage Daemons (OSDs) use IP addresses on the storage_mgmt network for cluster operations.
Changing the IP address of an OSD can cause the following issues:
-
The OSD to go into a
DOWNstate - The cluster to start data rebalancing to compensate for missing placement groups (PGs)
- Degraded performance during replication and data movement
- Potential data availability issues
While mitigation strategies exist to manage IP address changes, these operations are complex and carry risks. Plan your network addressing carefully during initial deployment to avoid the need for changes.
1.3.2.2. Co-locating services in a HCI environment for NUMA nodes Copy linkLink copied to clipboard!
A two-NUMA node system can host a latency sensitive Compute service workload on one NUMA node and a Ceph OSD workload on the other NUMA node. To configure Ceph OSDs to use a specific NUMA node not being used by the the Compute service, use either of the following Ceph OSD configurations:
-
osd_numa_nodesets affinity to a NUMA node (-1 for none). -
osd_numa_auto_affinityautomatically sets affinity to the NUMA node where storage and network match.
If there are network interfaces on both NUMA nodes and the disk controllers are on NUMA node 0, do the following:
- Use a network interface on NUMA node 0 for the storage network
- Host the Ceph OSD workload on NUMA node 0.
- Host the Compute service workload on NUMA node 1 and have it use the network interfaces on NUMA node 1.
Set osd_numa_auto_affinity to true, as in the initial Ceph configuration file. Alternatively, set the osd_numa_node directly to 0 and clear the osd_numa_auto_affinity parameter so that it defaults to false.
When a hyperconverged cluster backfills as a result of an OSD going offline, the backfill process can be slowed down. In exchange for a slower recovery, the backfill activity has less of an impact on the co-located Compute service (nova) workload. Red Hat Ceph Storage has the following defaults to control the rate of backfill activity.
-
osd_recovery_op_priority = 3 -
osd_max_backfills = 1 -
osd_recovery_max_active_hdd = 3 -
osd_recovery_max_active_ssd = 10
1.3.3. Confirming Red Hat Ceph Storage deployment Copy linkLink copied to clipboard!
Confirm Red Hat Ceph Storage is deployed before proceeding.
Procedure
- Connect to a data plane node by using SSH.
View the status of the Red Hat Ceph Storage cluster:
cephadm shell -- ceph -s
$ cephadm shell -- ceph -sCopy to Clipboard Copied! Toggle word wrap Toggle overflow
1.3.4. Confirming Red Hat Ceph Storage tuning Copy linkLink copied to clipboard!
Ensure that Red Hat Ceph Storage is properly tuned before proceeding.
Procedure
- Connect to a data plane node by using SSH.
Verify overall Red Hat Ceph Storage tuning with the following commands:
ceph config dump | grep numa ceph config dump | grep autotune ceph config dump | get mgr
$ ceph config dump | grep numa $ ceph config dump | grep autotune $ ceph config dump | get mgrCopy to Clipboard Copied! Toggle word wrap Toggle overflow Verify the tuning of an OSD with the following commands:
ceph config get <osd_number> osd_memory_target ceph config get <osd_number> osd_memory_target_autotune ceph config get <osd_number> osd_numa_auto_affinity
$ ceph config get <osd_number> osd_memory_target $ ceph config get <osd_number> osd_memory_target_autotune $ ceph config get <osd_number> osd_numa_auto_affinityCopy to Clipboard Copied! Toggle word wrap Toggle overflow -
Replace
<osd_number>with the number of an OSD. For example, to refer to OSD 11, useosd.11.
-
Replace
Verify the default backfill values of an OSD with the following commands:
ceph config get <osd_number> osd_recovery_op_priority ceph config get <osd_number> osd_max_backfills ceph config get <osd_number> osd_recovery_max_active_hdd ceph config get <osd_number> osd_recovery_max_active_ssd
$ ceph config get <osd_number> osd_recovery_op_priority $ ceph config get <osd_number> osd_max_backfills $ ceph config get <osd_number> osd_recovery_max_active_hdd $ ceph config get <osd_number> osd_recovery_max_active_ssdCopy to Clipboard Copied! Toggle word wrap Toggle overflow -
Replace
<osd_number>with the number of an OSD. For example, to refer to OSD 11, useosd.11.
-
Replace
1.4. Configuring the data plane to use the co-located Red Hat Ceph Storage server Copy linkLink copied to clipboard!
Although the Red Hat Ceph Storage cluster is physically co-located with the Compute services on the data plane nodes, it is treated as logically separated. Red Hat Ceph Storage must be configured as the storage solution before the data plane nodes can use it.
Prerequisites
- Complete the procedures in Integrating Red Hat Ceph Storage.
Procedure
-
Edit the
OpenStackDataPlaneNodeSetCR. To define the
cephxkey and configuration file for the Compute service (nova), use theextraMountsparameter.The following is an example of using the
extraMountsparameter for this purpose:Copy to Clipboard Copied! Toggle word wrap Toggle overflow -
Locate the
serviceslist in the CR. Edit the
serviceslist to restore all of the services removed in Configuring the data plane node networks. Restoring the fullserviceslist allows the remaining jobs to be run that complete the configuration of the HCI environment:Copy to Clipboard Copied! Toggle word wrap Toggle overflow NoteIn addition to restoring the default service list, the
ceph-clientservice is added after therun-osservice. Theceph-clientservice configures EDPM nodes as clients of a Red Hat Ceph Storage server. This service distributes the files necessary for the clients to connect to the Red Hat Ceph Storage server.Create a
ConfigMapto set thereserved_host_memory_mbparameter to a value appropriate for your configuration.The following is an example of a ConfigMap used for this purpose:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow NoteThe value for the
reserved_host_memory_mbparameter may be set so that the Compute service scheduler does not give memory to a virtual machine that a Ceph OSD on the same server needs. The example reserves 5 GB per OSD for 10 OSDs per host in addition to the default reserved memory for the hypervisor. In an IOPS-optimized cluster, you can improve performance by reserving more memory for each OSD. The 5 GB number is provided as a starting point which can be further tuned if necessary.Add
reserved-memory-novato theconfigMapslist by editing theOpenStackDataPlaneService/nova-custom-cephfile:Copy to Clipboard Copied! Toggle word wrap Toggle overflow Apply the CR changes.
oc apply -f <dataplane_cr_file>
$ oc apply -f <dataplane_cr_file>Copy to Clipboard Copied! Toggle word wrap Toggle overflow Replace
<dataplane_cr_file>with the name of your file.NoteAnsible does not configure or validate the networks until the
OpenStackDataPlaneDeploymentCRD is created.
-
Create an
OpenStackDataPlaneDeploymentCRD, as described in Creating the data plane in the Deploying Red Hat OpenStack Services on OpenShift guide, which has theOpenStackDataPlaneNodeSetCRD file defined above to have Ansible configure the services on the data plane nodes.