Maintaining the Red Hat OpenStack Services on OpenShift deployment
Maintaining a Red Hat OpenStack Services on OpenShift environment on a Red Hat OpenShift Container Platform cluster
Abstract
Providing feedback on Red Hat documentation Copy linkLink copied to clipboard!
We appreciate your feedback. Tell us how we can improve the documentation.
To provide documentation feedback for Red Hat OpenStack Services on OpenShift (RHOSO), create a Jira issue in the OSPRH Jira project.
Procedure
- Log in to the Red Hat Atlassian Jira.
- Click the following link to open a Create Issue page: Create issue
- Complete the Summary and Description fields. In the Description field, include the documentation URL, chapter or section number, and a detailed description of the issue.
- Click Create.
- Review the details of the bug you created.
Chapter 1. Accessing the RHOSO cloud Copy linkLink copied to clipboard!
You can access your Red Hat OpenStack Services on OpenShift (RHOSO) cloud to perform actions on your data plane by either accessing the OpenStackClient pod through a remote shell from your workstation, or by using a browser to access the Dashboard service (horizon) interface.
1.1. Accessing the OpenStackClient pod Copy linkLink copied to clipboard!
You can execute Red Hat OpenStack Services on OpenShift (RHOSO) commands on the deployed data plane by using the OpenStackClient pod through a remote shell from your workstation. The OpenStack Operator created the OpenStackClient pod as a part of the OpenStackControlPlane resource. The OpenStackClient pod contains the client tools and authentication details that you require to perform actions on your data plane.
Prerequisites
-
You are logged on to a workstation that has access to the Red Hat OpenShift Container Platform (RHOCP) cluster as a user with
cluster-adminprivileges.
Procedure
Access the remote shell for the
OpenStackClientpod:$ oc rsh -n openstack openstackclientRun your
openstackcommands. For example, you can create adefaultnetwork with the following command:$ openstack network create defaultExit the
OpenStackClientpod:$ exit
Additional resources
1.2. Accessing the Dashboard service (horizon) interface Copy linkLink copied to clipboard!
You can access the OpenStack Dashboard service (horizon) interface by providing the Dashboard service endpoint URL in a browser.
Prerequisites
- The Dashboard service is enabled on the control plane. For information about how to enable the Dashboard service, see Enabling the Dashboard service (horizon) interface in Customizing the Red Hat OpenStack Services on OpenShift deployment.
- You need to log into the Dashboard as the admin user.
Procedure
Retrieve the admin password from the
AdminPasswordparameter in theosp-secretsecret:$ oc get secret osp-secret -o jsonpath='{.data.AdminPassword}' | base64 -dRetrieve the Dashboard service endpoint URL:
$ oc get horizons horizon -o jsonpath='{.status.endpoint}'- Open a browser.
- Enter the Dashboard endpoint URL.
-
Log in to the Dashboard by providing the username of
adminand the admin password.
Chapter 2. Scaling data plane nodes Copy linkLink copied to clipboard!
You can scale out your data plane by adding new nodes to existing node sets and by adding new node sets. You can scale in your data plane by removing nodes from node sets and by removing node sets.
2.1. Adding nodes to a node set Copy linkLink copied to clipboard!
You can scale out your data plane by adding new nodes to the nodes section of an existing OpenStackDataPlaneNodeSet custom resource (CR).
Prerequisites
-
If you are adding unprovisioned nodes, then a
BareMetalHostCR must be registered and inspected for each bare-metal data plane node. Each bare-metal node must be in theAvailablestate after inspection.
Procedure
-
Open the
OpenStackDataPlaneNodeSetCR definition file for the node set you want to update, for example,openstack_data_plane.yaml. Add the new node to the node set:
Pre-Provisioned:
apiVersion: dataplane.openstack.org/v1beta1 kind: OpenStackDataPlaneNodeSet metadata: name: openstack-node-set spec: preProvisioned: True nodes: ... edpm-compute-2: hostName: edpm-compute-2 ansible: ansibleHost: 192.168.122.102 networks: - name: ctlplane subnetName: subnet1 defaultRoute: true fixedIP: 192.168.122.102 - name: internalapi subnetName: subnet1 - name: storage subnetName: subnet1 - name: tenant subnetName: subnet1 ...Unprovisioned:
apiVersion: dataplane.openstack.org/v1beta1 kind: OpenStackDataPlaneNodeSet metadata: name: openstack-node-set spec: preProvisioned: False nodes: ... edpm-compute-2: hostName: edpm-compute-2 ...
For information about the properties you can use to configure common node attributes, see OpenStackDataPlaneNodeSet CR spec properties in Deploying Red Hat OpenStack Services on OpenShift.
-
Save the
OpenStackDataPlaneNodeSetCR definition file. Apply the updated
OpenStackDataPlaneNodeSetCR configuration:$ oc apply -f openstack_data_plane.yamlVerify that the data plane resource has been updated by confirming that the status is
SetupReady:$ oc wait openstackdataplanenodeset openstack-node-set --for condition=SetupReady --timeout=10mWhen the status is
SetupReady, the command returns acondition metmessage, otherwise it returns a timeout error. For information about the data plane conditions and states, see Data plane conditions and states in Deploying Red Hat OpenStack Services on OpenShift.Create a file on your workstation to define the
OpenStackDataPlaneDeploymentCR:apiVersion: dataplane.openstack.org/v1beta1 kind: OpenStackDataPlaneDeployment metadata: name: <node_set_deployment_name>-
Replace
<node_set_deployment_name>with the name of theOpenStackDataPlaneDeploymentCR. The name must be unique, must consist of lower case alphanumeric characters,-(hyphen) or.(period), and must start and end with an alphanumeric character.
TipGive the definition file and the
OpenStackDataPlaneDeploymentCR unique and descriptive names that indicate the purpose of the modified node set.-
Replace
Add the
OpenStackDataPlaneNodeSetCR that you modified:spec: nodeSets: - <nodeSet_name>-
Save the
OpenStackDataPlaneDeploymentCR deployment file. Deploy the modified
OpenStackDataPlaneNodeSetCR:$ oc create -f openstack_data_plane_deploy.yaml -n openstackYou can view the Ansible logs while the deployment executes:
$ oc get pod -l app=openstackansibleee -w $ oc logs -l app=openstackansibleee -f --max-log-requests 10If the
oc logscommand returns an error similar to the following error, increase the--max-log-requestsvalue:error: you are attempting to follow 19 log streams, but maximum allowed concurrency is 10, use --max-log-requests to increase the limitVerify that the modified
OpenStackDataPlaneNodeSetCR is deployed:$ oc get openstackdataplanedeployment -n openstack NAME NODESETS STATUS MESSAGE openstack-data-plane ["openstack-data-plane"] True Setup Complete $ oc get openstackdataplanenodeset -n openstack NAME STATUS MESSAGE openstack-data-plane True NodeSet ReadyFor information about the meaning of the returned status, see Data plane conditions and states in Deploying Red Hat OpenStack Services on OpenShift.
If the status indicates that the data plane has not been deployed, then troubleshoot the deployment. For information, see Troubleshooting the data plane creation and deployment in Deploying Red Hat OpenStack Services on OpenShift.
If the new nodes are Compute nodes, you must bring them online:
Map the Compute nodes to the Compute cell that they are connected to:
$ oc rsh nova-cell0-conductor-0 nova-manage cell_v2 discover_hosts --verboseIf you did not create additional cells, this command maps the Compute nodes to
cell1.Verify that the hypervisor hostname is a fully qualified domain name (FQDN):
$ hostname -fIf the hypervisor hostname is not an FQDN, for example, if it was registered as a short name or full name instead, contact Red Hat Support.
Access the remote shell for the
openstackclientpod and verify that the deployed Compute nodes are visible on the control plane:$ oc rsh -n openstack openstackclient $ openstack hypervisor list
2.2. Adding a new node set to the data plane Copy linkLink copied to clipboard!
You can scale out your data plane by adding a new OpenStackDataPlaneNodeSet CR to the data plane. To add the new node set to an existing data plane, you must create a new OpenStackDataPlaneDeployment CR that deploys the new OpenStackDataPlaneNodeSet CR. If you want to be able to perform move operations, such as instance migration and resize, between your new node set and other node sets on your data plane, then you must also create an additional OpenStackDataPlaneDeployment CR that runs the ssh-known-hosts service on all the node sets between which move operations must be able to be performed.
Procedure
-
Create a file on your workstation to define the new
OpenStackDataPlaneNodeSetCR. Define the node set. For details about how to create a node set, see one of the following procedures:
Create a file on your workstation to define the
OpenStackDataPlaneDeploymentCR to deploy the new node set:apiVersion: dataplane.openstack.org/v1beta1 kind: OpenStackDataPlaneDeployment metadata: name: <node_set_deployment_name>-
Replace
<node_set_deployment_name>with the name of theOpenStackDataPlaneDeploymentCR. The name must be unique, must consist of lower case alphanumeric characters,-(hyphen) or.(period), and must start and end with an alphanumeric character.
TipGive the definition file and the
OpenStackDataPlaneDeploymentCR unique and descriptive names that indicate the purpose of the new node set.-
Replace
Add your new
OpenStackDataPlaneNodeSetCR to the list of node sets to deploy:spec: nodeSets: - <nodeSet_name>-
Save the
OpenStackDataPlaneDeploymentCR deployment file. Deploy the new
OpenStackDataPlaneNodeSetCR:$ oc create -f openstack_data_plane_deploy.yaml -n openstackYou can view the Ansible logs while the deployment executes:
$ oc get pod -l app=openstackansibleee -w $ oc logs -l app=openstackansibleee -f --max-log-requests 10If the
oc logscommand returns an error similar to the following error, increase the--max-log-requestsvalue:error: you are attempting to follow 19 log streams, but maximum allowed concurrency is 10, use --max-log-requests to increase the limitVerify that the new
OpenStackDataPlaneNodeSetCR is deployed:$ oc get openstackdataplanedeployment -n openstack NAME NODESETS STATUS MESSAGE openstack-data-plane ["openstack-data-plane"] True Setup Complete $ oc get openstackdataplanenodeset -n openstack NAME STATUS MESSAGE openstack-data-plane True NodeSet ReadyFor information about the meaning of the returned status, see Data plane conditions and states in Deploying Red Hat OpenStack Services on OpenShift.
If the status indicates that the data plane has not been deployed, then troubleshoot the deployment. For information, see Troubleshooting the data plane creation and deployment in Deploying Red Hat OpenStack Services on OpenShift.
If you want to be able to migrate workloads between your new node set and other node sets on your data plane, or perform resize operations, then you must create an additional
OpenStackDataPlaneDeploymentCR that runs thessh-known-hostsservice on all the node sets that the move operations are expected to work between:Create a file on your workstation to define an
OpenStackDataPlaneDeploymentCR that enables move operations:apiVersion: dataplane.openstack.org/v1beta1 kind: OpenStackDataPlaneDeployment metadata: name: enable-move-operationsAdd the new
OpenStackDataPlaneNodeSetCR to the list of node sets and all the existingOpenStackDataPlaneNodeSetCRs that move operations must be able to be performed between:spec: nodeSets: - enable-move-operations - ... - <nodeSet_name>Specify that only the
ssh-known-hostsservice is executed on the specified node sets when deploying the node sets in the move operationsOpenStackDataPlaneDeploymentCR:spec: ... servicesOverride: - ssh-known-hosts-
Save the
OpenStackDataPlaneDeploymentCR deployment file. Deploy the
ssh-known-hostsservice to enable move operations between the new node set and the other specified node sets on the data plane:$ oc create -f enable_move_operations.yaml -n openstack
If the new nodes are Compute nodes, you must bring them online:
Map the Compute nodes to the Compute cell that they are connected to:
$ oc rsh nova-cell0-conductor-0 nova-manage cell_v2 discover_hosts --verboseIf you did not create additional cells, this command maps the Compute nodes to
cell1.Verify that the hypervisor hostname is a fully qualified domain name (FQDN):
$ hostname -fIf the hypervisor hostname is not an FQDN, for example, if it was registered as a short name or full name instead, contact Red Hat Support.
Access the remote shell for the
openstackclientpod and verify that the deployed Compute nodes are visible on the control plane:$ oc rsh -n openstack openstackclient $ openstack hypervisor list
2.3. Removing a Compute node from the data plane Copy linkLink copied to clipboard!
You can remove a Compute node from a node set on the data plane. If you remove all the nodes from a node set, then you must also remove the node set from the data plane.
Prerequisites
-
You are logged in to the RHOCP cluster as a user with
cluster-adminprivileges. - The workloads on the Compute nodes have been migrated to other Compute nodes.
Procedure
Access the remote shell for the
openstackclientpod:$ oc rsh -n openstack openstackclientRetrieve the IP address of the Compute node that you want to remove:
$ openstack hypervisor listRetrieve a list of your Compute nodes to identify the name and UUID of the node that you want to remove:
$ openstack compute service listDisable the
nova-computeservice on the Compute node to be removed:$ openstack compute service set <hostname> nova-compute --disableTipUse the
--disable-reasonoption to add a short explanation on why the service is being disabled. This is useful if you intend to redeploy the Compute service.Exit the
OpenStackClientpod:$ exitSSH into the Compute node to be removed and stop the
ovnandnova-computecontainers:$ ssh -i <key_file_name> cloud-admin@<node_IP_address> [cloud-admin@<hostname> ~]$ sudo systemctl stop edpm_ovn_controller [cloud-admin@<hostname> ~]$ sudo systemctl stop edpm_ovn_metadata_agent [cloud-admin@<hostname> ~]$ sudo systemctl stop edpm_nova_compute-
Replace
<key_file_name>with the name and location of the SSH key pair file you created to enable Ansible to manage the RHEL nodes. -
Replace
<node_IP_address>with the IP address for the Compute node that you retrieved in step 2.
-
Replace
Remove the
systemdunit files that manage theovnandnova-computecontainers to prevent the agents from being automatically started and registered in the database if the removed node is rebooted:[cloud-admin@<hostname> ~]$ sudo rm -f /etc/systemd/system/edpm_ovn_controller [cloud-admin@<hostname> ~]$ sudo rm -f /etc/systemd/system/edpm_ovn_metadata_agent [cloud-admin@<hostname> ~]$ sudo rm -f /etc/systemd/system/edpm_nova_computeDisconnect from the Compute node:
$ exitAccess the remote shell for
openstackclient:$ oc rsh -n openstack openstackclientDelete the network agents for the Compute node to be removed:
$ openstack network agent list [--host <hostname>] $ openstack network agent delete <agent_id>Delete the
nova-computeservice for the Compute node to be removed:$ openstack compute service delete <node_uuid>-
Replace
<node_uuid>with the UUID of the node to be removed that you retrieved in step 3.
-
Replace
Exit the
OpenStackClientpod:$ exitRemove the node from the
OpenStackDataPlaneNodeSetCR:$ oc patch openstackdataplanenodeset/<node_set_name> --type json --patch '[{ "op": "remove", "path": "/spec/nodes/<node_name>" }]'-
Replace
<node_set_name>with the name of theOpenStackDataPlaneNodeSetCR that the node belongs to. -
Replace
<node_name>with the name of the node defined in thenodessection of theOpenStackDataPlaneNodeSetCR.
-
Replace
Create a file on your workstation to define the
OpenStackDataPlaneDeploymentCR to update the node set with the Compute node removed:apiVersion: dataplane.openstack.org/v1beta1 kind: OpenStackDataPlaneDeployment metadata: name: <node_set_deployment_name>-
Replace
<node_set_deployment_name>with the name of theOpenStackDataPlaneDeploymentCR. The name must be unique, must consist of lower case alphanumeric characters,-(hyphen) or.(period), and must start and end with an alphanumeric character.
TipGive the definition file and the
OpenStackDataPlaneDeploymentCR unique and descriptive names that indicate the purpose of the modified node set.-
Replace
Add the
OpenStackDataPlaneNodeSetCR that you removed the node from:spec: nodeSets: - <nodeSet_name>-
Save the
OpenStackDataPlaneDeploymentCR deployment file. Deploy the
OpenStackDataPlaneDeploymentCR to delete the removed nodes:$ oc create -f openstack_data_plane_deploy.yaml -n openstackYou can view the Ansible logs while the deployment executes:
$ oc get pod -l app=openstackansibleee -w $ oc logs -l app=openstackansibleee -f --max-log-requests 10If the
oc logscommand returns an error similar to the following error, increase the--max-log-requestsvalue:error: you are attempting to follow 19 log streams, but maximum allowed concurrency is 10, use --max-log-requests to increase the limitVerify that the modified
OpenStackDataPlaneNodeSetCR is deployed:$ oc get openstackdataplanedeployment -n openstack NAME NODESETS STATUS MESSAGE openstack-data-plane ["openstack-data-plane"] True Setup Complete $ oc get openstackdataplanenodeset -n openstack NAME STATUS MESSAGE openstack-data-plane True NodeSet ReadyFor information about the meaning of the returned status, see Data plane conditions and states in Deploying Red Hat OpenStack Services on OpenShift.
If the status indicates that the data plane has not been deployed, then troubleshoot the deployment. For information, see Troubleshooting the data plane creation and deployment in Deploying Red Hat OpenStack Services on OpenShift.
2.4. Removing an OpenStackDataPlaneNodeSet resource Copy linkLink copied to clipboard!
You can remove a whole node set from the data plane. To remove an OpenStackDataPlaneNodeSet resource, you must perform the following tasks:
-
Stop the
ovnandnova-computecontainers running on each Compute node in the node set. -
Disable and delete the
nova-computeservice from each Compute node in the node set. - Delete the network agent from each Compute node in the node set.
- Remove the SSH host keys of the removed nodes from the nodes in the remaining node sets.
- Delete the node set and remove the node set from the data plane.
Prerequisites
-
You are logged in to the RHOCP cluster as a user with
cluster-adminprivileges. - The workloads on the node set Compute nodes have been migrated to Compute nodes on another node set.
Procedure
Access the remote shell for the
openstackclientpod:$ oc rsh -n openstack openstackclientRetrieve the IP address of each Compute node you want to remove:
$ openstack hypervisor listRetrieve a list of your Compute nodes to identify the name and UUID of each node that you want to remove:
$ openstack compute service listDisable the
nova-computeservice on each Compute node to be removed:$ openstack compute service set <hostname> nova-compute --disableTipUse the
--disable-reasonoption to add a short explanation on why the service is being disabled. This is useful if you intend to redeploy the Compute service.Exit the
OpenStackClientpod:$ exitPerform the following operations on each Compute node to be removed:
SSH into the Compute node to be removed:
$ ssh -i <key_file_name> cloud-admin@<node_IP_address>-
Replace
<key_file_name>with the name and location of the SSH key pair file you created to enable Ansible to manage the RHEL nodes. -
Replace
<node_IP_address>with the IP address for the Compute node that you retrieved in step 2.
-
Replace
Stop the
ovnandnova-computecontainers:[cloud-admin@<hostname> ~]$ sudo systemctl stop edpm_ovn_controller [cloud-admin@<hostname> ~]$ sudo systemctl stop edpm_ovn_metadata_agent [cloud-admin@<hostname> ~]$ sudo systemctl stop edpm_nova_computeRemove the
systemdunit files that manage theovnandnova-computecontainers to prevent the agents from being automatically started and registered in the database if the removed node is rebooted:[cloud-admin@<hostname> ~]$ sudo rm -f /etc/systemd/system/edpm_ovn_controller [cloud-admin@<hostname> ~]$ sudo rm -f /etc/systemd/system/edpm_ovn_metadata_agent [cloud-admin@<hostname> ~]$ sudo rm -f /etc/systemd/system/edpm_nova_computeDisconnect from the Compute node:
$ exit
Access the remote shell for
openstackclient:$ oc rsh -n openstack openstackclientDelete the network agents for each Compute node in the node set:
$ openstack network agent list [--host <hostname>] $ openstack network agent delete <agent_id>Delete the
nova-computeservice for each Compute node in the node set:$ openstack compute service delete <node_uuid>-
Replace
<node_uuid>with the UUID of the node to be removed that you retrieved in step 3.
-
Replace
Exit the
openstackclientpod:$ exitSearch for the string
secretHashesin the output of the following command for the secrets in the node set to be deleted:$ oc get openstackdataplanenodeset <node_set_name> -n openstack -o yamlThe
secretHashesfield lists all the node set secrets in key-value pair format:<key>:<value>. The following example illustrates thesecretHashesformat in YAML output:secretHashes: cert-libvirt-default-compute-4drna21w-0: n68chbfh678h5dfhcfh576h546h566h5c4h5cdh679hffh67h79h98h56fhc4h588h58fhb4h548h59fh554h54fh5cdh646h577hffhbdh569h5f9h68bq cert-libvirt-default-compute-4drna21w-1: n68dhdbh65chdbh5f7h695h7h54chcdh654h59ch564h5d6hdch66h54bh66ch556h649h666h76h55hc7h564h65dh5fch5c7h5fbh8bh55hcbh5b5qDelete the node set:
$ oc delete openstackdataplanenodeset/<node_set_name> -n openstack-
Replace
<node_set_name>with the name of theOpenStackDataPlaneNodeSetCR to be deleted.
-
Replace
Delete the node set secrets:
$ oc delete secret <secret_name>-
Replace
<secret_name>with the key of thesecretHasheskey-value pair you retrieved in the previous step, for example,cert-libvirt-default-compute-4drna21w-0.
NoteYou can ensure that secrets created by
cert-managerget removed automatically by setting the--enable-certificate-owner-refflag for the cert-manager Operator for Red Hat OpenShift. For more information, see Deleting a TLS secret automatically upon Certificate removal.-
Replace
If the node set you removed listed the global
ssh-known-hostsservice, then you must add thessh-known-hostsservice to one of the remainingOpenStackDataPlaneNodeSetCRs listed in theOpenStackDataPlaneDeploymentCR. Open the definition file for one of the remainingOpenStackDataPlaneNodeSetCRs from your workstation and add thessh-known-hostsservice to theservicesfield in the order that it should be executed relative to the other services:apiVersion: dataplane.openstack.org/v1beta1 kind: OpenStackDataPlaneNodeSet metadata: name: <node_set_name> spec: services: - download-cache - bootstrap - configure-network - validate-network - install-os - configure-os - ssh-known-hosts - run-os - libvirt - nova - ovn - neutron-metadata - telemetryNoteWhen adding the
ssh-known-hostsservice to the services list in a node set definition, you must include all the required services, including the default services. If you include only thessh-known-hostsservice in the services list, then that is the only service that is deployed.-
Save the updated
OpenStackDataPlaneNodeSetCR definition file. Apply the updated
OpenStackDataPlaneNodeSetCR configuration:$ oc apply -f <node_set_name>.yamlCreate a file on your workstation to define the
OpenStackDataPlaneDeploymentCR that removes the node set from the data plane:apiVersion: dataplane.openstack.org/v1beta1 kind: OpenStackDataPlaneDeployment metadata: name: <node_set_deployment_name>-
Replace
<node_set_deployment_name>with the name of theOpenStackDataPlaneDeploymentCR. The name must be unique, must consist of lower case alphanumeric characters,-(hyphen) or.(period), and must start and end with an alphanumeric character.
TipGive the definition file and the
OpenStackDataPlaneDeploymentCR unique and descriptive names that indicate the purpose of the modified node set.-
Replace
Add the remaining
OpenStackDataPlaneNodeSetCRs in the data plane to the list of node sets to deploy:spec: nodeSets: - <nodeSet_name>Specify that the
OpenStackDataPlaneDeploymentCR should only run thessh-known-hostsservice when deploying the listed node sets:spec: ... servicesOverride: - ssh-known-hosts-
Save the
OpenStackDataPlaneDeploymentCR deployment file. Deploy the
ssh-known-hostsservice to delete the removed nodes from the known hosts lists on the remaining nodes:$ oc create -f openstack_data_plane_deploy.yaml -n openstackYou can view the Ansible logs while the deployment executes:
$ oc get pod -l app=openstackansibleee -w $ oc logs -l app=openstackansibleee -f --max-log-requests 10If the
oc logscommand returns an error similar to the following error, increase the--max-log-requestsvalue:error: you are attempting to follow 19 log streams, but maximum allowed concurrency is 10, use --max-log-requests to increase the limitVerify that the modified
OpenStackDataPlaneNodeSetCR is deployed:$ oc get openstackdataplanedeployment -n openstack NAME NODESETS STATUS MESSAGE openstack-data-plane ["openstack-data-plane"] True Setup Complete $ oc get openstackdataplanenodeset -n openstack NAME STATUS MESSAGE openstack-data-plane True NodeSet ReadyFor information about the meaning of the returned status, see Data plane conditions and states in Deploying Red Hat OpenStack Services on OpenShift.
If the status indicates that the data plane has not been deployed, then troubleshoot the deployment. For information, see Troubleshooting the data plane creation and deployment in Deploying Red Hat OpenStack Services on OpenShift.
Chapter 3. Replacing data plane nodes Copy linkLink copied to clipboard!
You can replace pre-provisioned and provisioned data plane nodes without scaling the Red Hat OpenStack Services on OpenShift (RHOSO) cloud.
3.1. Replacing a pre-provisioned node Copy linkLink copied to clipboard!
When you replace a faulty pre-provisioned node, your replacement node has the same hostname and IP address as the faulty node. You create a new OpenStackDataPlaneDeployment CR to deploy the new node.
If your replacement node has a different IP address to the faulty node, you set a new fixedIP for the control plane network for the node and a new ansibleHost for the node in the OpenStackDataPlaneNodeSet CR. Then you create a new OpenStackDataPlaneDeployment CR to deploy the new node.
When you replace a pre-provisioned node, you manually clean up the node you are removing.
For information about errors, data plane conditions and states, and returned statuses that can occur when you modify an OpenStackDataPlaneNodeSet CR, see Modifying an OpenStackDataPlaneNodeSet CR in Customizing the Red Hat OpenStack Services on OpenShift deployment.
Prerequisites
-
You are logged in to the RHOCP cluster as a user with
cluster-adminprivileges. - The workloads on the Compute nodes have been migrated to other Compute nodes.
Procedure
If the faulty node is still reachable, perform clean-up tasks to ensure that the faulty node does not impact the new node.
SSH into the removed node, and stop the
ovnandnova-computecontainers:$ ssh -i <key_file_name> cloud-admin@<node_IP_address> [cloud-admin@<hostname> ~]$ sudo systemctl stop edpm_ovn_controller [cloud-admin@<hostname> ~]$ sudo systemctl stop edpm_ovn_metadata_agent [cloud-admin@<hostname> ~]$ sudo systemctl stop edpm_nova_compute-
Replace
<key_file_name>with the name and location of the SSH key pair file you created to enable Ansible to manage the RHEL nodes. -
Replace
<node_IP_address>with the IP address for the removed node.
-
Replace
Remove the
systemdunit files that manage theovnandnova-computecontainers to prevent the agents from being automatically started and registered in the database if the removed node is rebooted:[cloud-admin@<hostname> ~]$ sudo rm -f /etc/systemd/system/edpm_ovn_controller [cloud-admin@<hostname> ~]$ sudo rm -f /etc/systemd/system/edpm_ovn_metadata_agent [cloud-admin@<hostname> ~]$ sudo rm -f /etc/systemd/system/edpm_nova_computeDisconnect from the node:
$ exit
If your replacement node has the same IP address as the faulty node, proceed to Step 3 to create the
OpenStackDataPlaneDeploymentCR to deploy the new node. If your replacement node has a different IP address to the faulty node, open theOpenStackDataPlaneNodeSetCR file for the node set you want to update, for example,openstack_data_plane.yaml, and make the following changes.Update the
fixedIPfor the control plane network for the node you are replacing to specify the IP address of the new node:Example:
nodes: edpm-compute-0: hostName: edpm-compute-0 networks: - name: ctlplane subnetName: subnet1 defaultRoute: true fixedIP: 192.168.122.100 ...Update the
ansibleHostvalue for the node you are replacing to specify the IP address of the new node:Example:
nodes: edpm-compute-0: hostName: edpm-compute-0 ... ansible: ansibleHost: 192.168.122.100 ansibleUser: cloud-admin ansibleVars: fqdn_internal_api: edpm-compute-0.example.com ...-
Save the
OpenStackDataPlaneNodeSetCR file. Apply the updated
OpenStackDataPlaneNodeSetCR configuration:$ oc apply -f openstack_data_plane.yamlVerify that the data plane resource has been updated by confirming that the status is
SetupReady:$ oc wait openstackdataplanenodeset openstack-data-plane --for condition=SetupReady --timeout=10mWhen the status is
SetupReady, the command returns acondition metmessage, otherwise it returns a timeout error.
Create a file on your workstation to define the
OpenStackDataPlaneDeploymentCR to deploy the new node:apiVersion: dataplane.openstack.org/v1beta1 kind: OpenStackDataPlaneDeployment metadata: name: <node_set_deployment_name>-
Replace
<node_set_deployment_name>with the name of theOpenStackDataPlaneDeploymentCR. The name must be unique, must consist of lower case alphanumeric characters,-(hyphen) or.(period), and must start and end with an alphanumeric character.
-
Replace
Add the
OpenStackDataPlaneNodeSetCR:spec: nodeSets: - <nodeSet_name>-
Save the
OpenStackDataPlaneDeploymentCR deployment file. Deploy the
OpenStackDataPlaneNodeSetCR:$ oc create -f openstack_data_plane_deploy.yaml -n openstackYou can view the Ansible logs while the deployment executes:
$ oc get pod -l app=openstackansibleee -w $ oc logs -l app=openstackansibleee -f --max-log-requests 10Verify that the
OpenStackDataPlaneNodeSetCR is deployed:$ oc get openstackdataplanedeployment -n openstack NAME STATUS MESSAGE openstack-data-plane True Setup Complete $ oc get openstackdataplanenodeset -n openstack NAME STATUS MESSAGE openstack-data-plane True NodeSet Ready
3.2. Replacing a provisioned node Copy linkLink copied to clipboard!
To replace a faulty provisioned node from the data plane without scaling the Red Hat OpenStack Services on OpenShift (RHOSO) cloud, you delete the faulty bare metal host (BMH). The OpenStackBaremetalSet CR is reconciled to provision a new available BMH and reset the deployment status of the OpenStackDataPlaneNodeSet, prompting you to create a new OpenStackDataPlaneDeployment CR for deploying on the newly provisioned node.
Prerequisites
-
You are logged in to the RHOCP cluster as a user with
cluster-adminprivileges. - The workloads on the Compute nodes have been migrated to other Compute nodes.
Procedure
Verify that you have a spare BMH in the
availablestate that you can use to replace the faulty node:$ oc get bmhExample output:
NAME STATE CONSUMER ONLINE ERROR AGE leaf0-0 available false 11h leaf0-1 provisioned nodeset-0 true 11h leaf1-0 provisioned nodeset-1 true 11h leaf1-1 provisioned nodeset-1 true 11hDelete the faulty node:
$ oc delete bmh leaf0-1Example output:
baremetalhost.metal3.io "leaf0-1" deletedThe
OpenStackBaremetalSetCR is reconciled to provision a new available BMH and reset the deployment status of theOpenStackDataPlaneNodeSet, prompting you to create a newOpenStackDataPlaneDeploymentCR for deploying on the newly provisioned node.Wait for the node set that contained the faulty node to reach the
SetupReadystate.$ oc wait openstackdataplanenodeset openstack-data-plane --for condition=SetupReady --timeout=10mWhen the status is
SetupReady, the command returns acondition metmessage, otherwise it returns a timeout error.Create a file on your workstation to define the
OpenStackDataPlaneDeploymentCR:apiVersion: dataplane.openstack.org/v1beta1 kind: OpenStackDataPlaneDeployment metadata: name: <node_set_deployment_name>Replace
<node_set_deployment_name>with the name of theOpenStackDataPlaneDeploymentCR. The name must be unique, must consist of lower case alphanumeric characters,-(hyphen) or.(period), and must start and end with an alphanumeric character.TipGive the definition file and the
OpenStackDataPlaneDeploymentCR unique and descriptive names that indicate the purpose of the modified node set.
Add the
OpenStackDataPlaneNodeSetCR that you modified:spec: nodeSets: - <nodeSet_name>-
Save the
OpenStackDataPlaneDeploymentCR deployment file. Deploy the modified
OpenStackDataPlaneNodeSetCR:$ oc create -f openstack_data_plane_deploy.yaml -n openstackYou can view the Ansible logs while the deployment executes:
$ oc get pod -l app=openstackansibleee -w $ oc logs -l app=openstackansibleee -f --max-log-requests 10Verify that the modified
OpenStackDataPlaneNodeSetCR is deployed:$ oc get openstackdataplanedeployment -n openstack NAME STATUS MESSAGE openstack-data-plane True Setup Complete $ oc get openstackdataplanenodeset -n openstack NAME STATUS MESSAGE openstack-data-plane True NodeSet ReadyVerify that the spare BMH has replaced the faulty BMH:
$ oc get bmhExample output:
NAME STATE CONSUMER ONLINE ERROR AGE leaf0-0 provisioned nodeset-0 true 13h leaf1-0 provisioned nodeset-1 true 13h leaf1-1 provisioned nodeset-1 true 13h
Chapter 4. Rebooting data plane nodes Copy linkLink copied to clipboard!
You might need to reboot the nodes on your data plane. The reboot method is determined by the type of node. You can also configure how your data plane handles the reboot process based on the node type.
4.1. Configuring the data plane Compute node reboot behavior Copy linkLink copied to clipboard!
You can configure how the data plane handles the shut down and restart of instances that are hosted on Compute nodes if the instances are not migrated off the host node before the host Compute node is rebooted.
Prerequisites
-
You are logged on to a workstation that has access to the Red Hat OpenShift Container Platform (RHOCP) cluster as a user with
cluster-adminprivileges.
Procedure
Open the
nova-extra-config.yamldefinition file for the default Compute service (nova)ConfigMapcustom resource (CR) namednova-extra-configand add the following configuration:apiVersion: v1 kind: ConfigMap metadata: name: nova-extra-config namespace: openstack data: 30-nova-reboot.conf: | [DEFAULT] resume_guests_state_on_host_boot=true1 shutdown_timeout=3002 - 1
- Set to
trueto return instances to the same state on the Compute node after reboot. When set toFalse, the instances remain down and you must start them manually. - 2
- Specify the number of seconds to wait for an instance to perform a controlled, clean shut down before being powered off and rebooted. Do not set this value to
0. A value of 0 (zero) means that the instance is powered off immediately with no opportunity for instance OS clean-up. The default value is60.
Create a file on your workstation to define the
OpenStackDataPlaneDeploymentCR:apiVersion: dataplane.openstack.org/v1beta1 kind: OpenStackDataPlaneDeployment metadata: name: <node_set_deployment_name>-
Replace
<node_set_deployment_name>with the name of theOpenStackDataPlaneDeploymentCR. The name must be unique, must consist of lower case alphanumeric characters,-(hyphen) or.(period), and must start and end with an alphanumeric character.
TipGive the definition file and the
OpenStackDataPlaneDeploymentCR unique and descriptive names that reflect the purpose of the node sets in the deployment.-
Replace
Add the
OpenStackDataPlaneNodeSetCRs that configure the Compute nodes on the data plane that run the defaultnovaservice:spec: nodeSets: - <nodeSet_name>-
Replace
<nodeSet_name>with the names of theOpenStackDataPlaneNodeSetCRs that you want to include in your data plane deployment.
-
Replace
-
Save the
OpenStackDataPlaneDeploymentCR deployment file. Deploy the modified
OpenStackDataPlaneNodeSetCRs:$ oc create -f openstack_data_plane_deploy.yaml -n openstackYou can view the Ansible logs while the deployment executes:
$ oc get pod -l app=openstackansibleee -w $ oc logs -l app=openstackansibleee -f --max-log-requests 10Verify that the modified
OpenStackDataPlaneNodeSetCRs are deployed:$ oc get openstackdataplanedeployment -n openstack NAME STATUS MESSAGE openstack-data-plane True Setup Complete $ oc get openstackdataplanenodeset -n openstack NAME STATUS MESSAGE openstack-data-plane True NodeSet ReadyFor information about the meaning of the returned status, see Data plane conditions and states in the Deploying Red Hat OpenStack Services on OpenShift guide.
If the status indicates that the data plane has not been deployed, then troubleshoot the deployment. For information, see Troubleshooting the data plane creation and deployment in the Deploying Red Hat OpenStack Services on OpenShift guide.
4.2. Rebooting Object Storage service (swift) nodes Copy linkLink copied to clipboard!
To reboot Object Storage service (swift) nodes, complete the following steps for every Object Storage node in your cluster. Reboot nodes one after the other individually instead at the same time to ensure that the cluster remains available even when a node is rebooting.
Procedure
- Log in to an Object Storage node.
Reboot the node:
$ sudo reboot- Wait until the node boots.
- Repeat the reboot for each Object Storage node in the cluster.
4.3. Rebooting Compute nodes Copy linkLink copied to clipboard!
You might need to reboot the Compute nodes on the Red Hat OpenShift on OpenStack (RHOSO) data plane. You can migrate all the virtual machine instances hosted on the Compute nodes that you want to reboot to other hosts before performing the reboot, which is recommended to ensure minimal downtime of instance workloads in your RHOSO environment. If you do not migrate the instances off the host Compute nodes before reboot, then how the instances are restarted is based on how the data plane is configured to handle instances on host reboot. For more information about Compute node reboot behavior, see Configuring the data plane Compute node reboot behavior.
When you restart the Compute service (nova) and your Compute host detects a name change, you must discover why the Compute host name changed and correct that condition to return the hostname to the original value. When you correct the issue, you must restart the Compute service. For more information, see Section 4.3.1, “Troubleshooting Compute hostname change detection”.
If you have a Multi-RHEL environment, and you want to migrate virtual machines from a Compute node that is running RHEL 9.4 or 9.6 to a Compute node that is running RHEL 9.2, only cold migration is supported. For more information about cold migration, see Cold migrating an instance in Configuring the Compute service for instance creation.
Procedure
Access the remote shell for the
openstackclientpod:$ oc rsh -n openstack openstackclientRetrieve a list of your Compute nodes to identify the host name of the nodes that you want to reboot:
$ openstack compute service listDisable the Compute service on each Compute node that you want to reboot so that it does not provision new instances on the nodes:
$ openstack compute service set <hostname> nova-compute --disable-
Replace
<hostname>with the host name of the Compute node on which you are disabling the service.
-
Replace
List all the instances that are hosted on each Compute node:
$ openstack server list --host <hostname> --all-projectsOptional: Migrate the instances to another Compute node:
$ openstack server migrate --live-migration [--host <dest>] <instance> --wait-
Replace
<instance>with the name or ID of the instance to migrate. -
Optional: Include the
--hostoption and replace<dest>with the name or ID of the destination Compute node. If the--hostoption is not specified, then thenova-schedulerselects the destination node for the instances.
Repeat this migration command for each instance until none remain on the Compute node.
-
Replace
Exit the
openstackclientpod:$ exitCreate an
OpenStackDataPlaneDeploymentCR to reboot the nodes and save it to a file namedcompute_reboot_nodes_deploy.yamlon your workstation::apiVersion: dataplane.openstack.org/v1beta1 kind: OpenStackDataPlaneDeployment metadata: name: compute-node-reboot namespace: openstack spec: nodeSets:1 - <nodeSet_name> - ... - <nodeSet_name> servicesOverride: - reboot-os ansibleExtraVars: edpm_reboot_strategy: force ansibleLimit: <node_hostname>,...,<node_hostname>-
Replace
<nodeSet_name>with the names of theOpenStackDataPlaneNodeSetCRs that contain the nodes that you are rebooting. -
Replace
<node_hostname>with the names of the nodes in the node set to reboot. IfansibleLimitis not set, all the nodes in the node set are rebooted at the same time.
-
Replace
Deploy the data plane:
$ oc create -f compute_reboot_nodes_deploy.yaml -n openstackVerify that the
compute-node-rebootdeployment completed:$ oc get openstackdataplanedeployment NAME STATUS MESSAGE compute-node-reboot True Setup completeIf the deployment fails, see Troubleshooting data plane creation and deployment in the Deploying Red Hat OpenStack Services on OpenShift guide.
Access the remote shell for the
openstackclientpod:$ oc rsh -n openstack openstackclientRe-enable the Compute service on each rebooted Compute node:
$ openstack compute service set <hostname> nova-compute --enableConfirm that the Compute service on each Compute node is enabled:
$ openstack compute service listIf you did not migrate the instances that are hosted on the Compute node before reboot, then check that the instances have restarted and are in the correct state:
$ openstack server list --host <host> --all-projectsExit the
openstackclientpod:$ exit
4.3.1. Troubleshooting Compute hostname change detection Copy linkLink copied to clipboard!
If you start the Compute service (nova) and your Compute host detects a name change, you receive the following error message:
Found 7 instances on hypervisor but no ComputeNode record exists with host compute01.example.com node compute01.example.com; potential rename detected - aborting startup
To recover from this error, you must know the reason for the change of the hostname. Here are the three possible reasons for the change of the hostname:
- You have changed the name of the Compute host.
- The DNS is not working correctly and has assigned a different name to the Compute host.
- There is another deployment issue causing a Compute hostname to be incorrect.
When you resolve the issue, you must restart the Compute service.
The nova-libvirt container contains information about the hostname. If you resolve an issue with a Compute hostname, restart both the nova-libvirt and nova-compute containers to ensure that all components contain the correct name.
For further assistance, contact the Red Hat Technical Support Team.
Chapter 5. Removing a RHOSO deployment from the RHOCP environment Copy linkLink copied to clipboard!
You can remove a Red Hat OpenStack Services on OpenShift (RHOSO) deployment from the Red Hat OpenShift Container Platform (RHOCP) environment if you no longer require the RHOSO deployment. You can also remove any Operators that were installed on your RHOCP cluster for use only by the RHOSO deployment.
To create a new RHOSO deployment with the same data plane nodes after RHOSO is removed from the RHOCP environment, you must reprovision the nodes.
5.1. Removing RHOSO resources from the RHOCP environment Copy linkLink copied to clipboard!
You can remove the resources created for your Red Hat OpenStack Services on OpenShift (RHOSO) deployment from the Red Hat OpenShift Container Platform (RHOCP) environment if you no longer require the RHOSO deployment.
Prerequisites
-
You are logged on to a workstation that has access to the Red Hat OpenShift Container Platform (RHOCP) cluster as a user with
cluster-adminprivileges.
Procedure
Delete all the
OpenStackDataPlaneDeploymentobjects in the RHOSO namespace:$ oc delete OpenStackDataPlaneDeployment --all -n openstackDelete all the
OpenStackDataPlaneNodeSetobjects in the RHOSO namespace:$ oc delete OpenStackDataPlaneNodeSet --all -n openstackDelete all the
OpenStackDataPlaneServiceobjects in the RHOSO namespace:$ oc delete OpenStackDataPlaneService --all -n openstackDelete the
OpenStackControlPlaneobject from the RHOSO namespace:$ oc delete openstackcontrolplane \ -l core.openstack.org/openstackcontrolplane -n openstackConfirm that all the deployment pods are deleted:
$ oc get pods -n openstackOptional: Delete all Persistent Volume Claims (PVCs) from the RHOSO namespace:
$ oc delete --all PersistentVolumeClaim -n openstackTipCreate a backup of your PVCs before deletion if you need to reuse them. For information about how to create a PVC backup, see the documentation for the storage backend used in your RHOSO environment.
Optional: Release the PV to make it available for other applications:
$ oc patch PersistentVolume <pv_name> -p '{"spec":{"claimRef": null}}'NoteIf you intend to re-use the released PV, ensure that the PV is cleaned before use. For more information, see Lifecycle of a volume and claim in the RHOCP Storage guide.
Delete the
Secretobjects that contain the certificates issued for the control plane services and the data plane nodes when the RHOSO environment was deployed:$ oc delete secret -l service-cert -n openstack $ oc delete secret -l ca-cert -n openstack $ oc delete secret -l osdp-service -n openstackIf the above commands return the message "No resources found" then the certificate secrets were automatically deleted when you deleted the control plane and data plane resources. For information about configuring the cert-manager Operator before RHOSO deployment to automatically delete the certificate secrets, see Deleting a TLS secret automatically upon Certificate removal in the RHOCP Security and Compliance guide.
Verify that the resources have been deleted from the namespace:
$ oc get all No resources found.Optional: Delete the namespace:
$ oc delete namespace openstackNoteYou do not need to delete the namespace if you plan to create a new RHOSO deployment in the same namespace.
If deleting the namespace is stuck in the terminating state:
Check if any remaining objects have a finalizer:
$ oc get $(oc api-resources|grep openstack.org|cut -d" " -f1 |paste -sd "," -),all -o custom-columns=Kind:.kind,Name:.metadata.name,Finalizers:.metadata.finalizers -n <namespace>Repeat the following command for each remaining object to remove the object finalizers and unblock deletion of each object:
$ oc patch -n <namespace> <object-name> -p '{"metadata":{"finalizers":[]}}' --type=merge
5.2. Removing RHOSO Operators from the RHOCP cluster Copy linkLink copied to clipboard!
You can remove any Operators that were installed on your Red Hat OpenShift Container Platform (RHOCP) cluster for use only by the Red Hat OpenStack Services on OpenShift (RHOSO) deployment. For more information about how to remove an Operator and uninstall all the resources associated with the Operator, see Deleting Operators from a cluster in the RHOCP Operators guide.
Chapter 6. Shutting down and starting up RHOSO nodes Copy linkLink copied to clipboard!
To perform maintenance on your Red Hat OpenStack on OpenShift (RHOSO) environment, you must shut down and start up the Red Hat OpenShift Container Platform (RHOCP) cluster and all the data plane nodes in a specific order to ensure minimal issues when you restart your cluster and data plane nodes.
Prerequisites
- An operational RHOSO environment.
-
You are logged on to a workstation that has access to the RHOSO control plane as a user with
cluster-adminprivileges. -
The
occommand line tool is installed on the workstation.
6.1. RHOSO deployment shutdown order Copy linkLink copied to clipboard!
To shut down the Red Hat OpenStack on OpenShift (RHOSO) environment, you must shut down the instances that host the workload, the data plane nodes, and the Red Hat OpenShift Container Platform (RHOCP) cluster nodes in the following order:
- Shut down instances hosted on the Compute nodes on the data plane.
- If your data plane includes hyperconverged infrastructure (HCI) nodes, shut down the Red Hat Ceph Storage cluster.
- Shut down Compute nodes.
- Shut down the RHOCP cluster nodes.
6.2. Shutting down instances hosted on the Compute nodes Copy linkLink copied to clipboard!
To shut down the Red Hat OpenStack Services on OpenShift (RHOSO) environment, you must first shut down all instances hosted on Compute nodes before shutting down the Compute nodes.
Procedure
Access the remote shell for the
OpenStackClientpod from your workstation:$ oc rsh -n openstack openstackclientList all running instances:
$ openstack server list --all-projectsStop each instance:
$ openstack server stop <instance_UUID>Repeat this step for each instance until you stop all running instances.
Exit the
OpenStackClientpod:$ exit
6.3. Shutting down the Red Hat Ceph Storage cluster for HCI environments Copy linkLink copied to clipboard!
If your data plane includes hyperconverged infrastructure (HCI) nodes, shut down the Red Hat Ceph Storage cluster. For more information about how to shut down the Red Hat Ceph Storage cluster, see "Powering down and rebooting the cluster using the Ceph Orchestrator" in the Red Hat Ceph Storage Administration Guide:
6.4. Shutting down Compute nodes Copy linkLink copied to clipboard!
As a part of shutting down the Red Hat OpenStack Services on OpenShift (RHOSO) environment, log in to and shut down each Compute node. For information about safely powering off bare-metal hosts, see Powering off bare-metal hosts in the RHOCP Scalability and performance guide.
Prerequisites
- You have stopped all instances hosted on the Compute nodes.
Procedure
Retrieve a list of the Compute nodes:
$ oc rsh -n openstack openstackclient openstack compute service listIf your Compute nodes are bare-metal nodes that were provisioned when deploying the data plane, annotate the
BareMetalHostcustom resource (CR) for each Compute node to prevent the Cluster Baremetal Operator (CBO) from rebooting the node:$ oc annotate bmh <compute_node_name> -n openshift-machine-api 'baremetalhost.metal3.io/detached=""'Log in as the
rootuser to a Compute node and shut down the node:# shutdown -h nowRepeat this step for each Compute node until you shut down all Compute nodes.
Verify that the Compute nodes are shut down:
$ oc rsh -n openstack openstackclient openstack hypervisor list -c ID -c State +--------------------------------------+---------+ | ID | State | +--------------------------------------+---------+ | 756968fd-272e-48d2-8f4a-54ef772b2acb | down | +--------------------------------------+---------+
6.5. Shutting down the RHOCP cluster Copy linkLink copied to clipboard!
As a part of shutting down the Red Hat OpenStack Services on OpenShift (RHOSO) environment, you must shut down the Red Hat OpenShift Container Platform (RHOCP) cluster that hosts the RHOSO environment. For information about how to shut down a RHOCP cluster, see Shutting down the cluster gracefully in the RHOCP Backup and restore guide.
6.6. RHOSO deployment startup order Copy linkLink copied to clipboard!
To start the Red Hat OpenStack Services on OpenShift (RHOSO) environment, you must start the Red Hat OpenShift Container Platform (RHOCP) cluster and data plane nodes in the following order:
- Start the RHOCP cluster.
- If your data plane includes hyperconverged infrastructure (HCI) nodes, start up the Red Hat Ceph Storage cluster.
- Start Compute nodes.
- Start instances on the Compute nodes.
6.7. Starting the RHOCP cluster Copy linkLink copied to clipboard!
As a part of starting up the Red Hat OpenStack Services on OpenShift (RHOSO) environment, you must start the Red Hat OpenShift Container Platform (RHOCP) cluster that hosts the RHOSO environment. For information about how to start up a RHOCP cluster, see Restarting the cluster gracefully in the RHOCP Backup and restore guide.
6.8. Starting the Red Hat Ceph Storage cluster for HCI environments Copy linkLink copied to clipboard!
If your data plane includes hyperconverged infrastructure (HCI) nodes, you must use the cephadm utility to unset the noout, norecover, norebalance, nobackfill, and nodown properties, and pause flagsstart. For more information about how to start the Red Hat Ceph Storage cluster, see "Powering down and rebooting the cluster using the Ceph Orchestrator" in the Red Hat Ceph Storage Administration Guide:
6.9. Starting Compute nodes Copy linkLink copied to clipboard!
As a part of starting the Red Hat OpenStack Services on OpenShift (RHOSO) environment, power on each Compute node and check the services on the node.
Prerequisites
- Powered down Compute nodes.
Procedure
- Power on each Compute node.
Re-attach Compute nodes that you detached from the Cluster Baremetal Operator (CBO) during shut down:
$ oc annotate bmh <compute_node_name> -n openshift-machine-api 'baremetalhost.metal3.io/detached-'
Verification
-
Log in to each Compute as the
rootuser. Check the services on the Compute node:
$ systemctl -t service
6.10. Starting instances on Compute nodes Copy linkLink copied to clipboard!
As a part of starting the Red Hat OpenStack Services on OpenShift (RHOSO) environment, start the instances on the Compute nodes.
Procedure
Access the remote shell for the
OpenStackClientpod from your workstation:$ oc rsh -n openstack openstackclientList all the instances:
$ openstack server list --all-projectsStart an instance:
$ openstack server start <instance_UUID>Repeat this step for each instance until you start all the instances.
Exit the
OpenStackClientpod:$ exit