Keeping Red Hat OpenStack Platform Updated
Performing minor updates of Red Hat OpenStack Platform
Abstract
Chapter 1. Introduction Copy linkLink copied to clipboard!
This document provides a workflow to help keep your Red Hat OpenStack Platform 16.0 environment updated with the latest packages and containers.
This guide provides an upgrade path through the following versions:
| Old Overcloud Version | New Overcloud Version |
|---|---|
| Red Hat OpenStack Platform 16.0 | Red Hat OpenStack Platform 16.0.z |
1.1. High level workflow Copy linkLink copied to clipboard!
The following table provides an outline of the steps required for the upgrade process:
| Step | Description |
|---|---|
| Updating the undercloud | Update the undercloud to the latest OpenStack Platform 16.0.z version. |
| Updating the overcloud | Update the overcloud to the latest OpenStack Platform 16.0.z version. |
| Updating the Ceph Storage nodes | Upgrade all Ceph Storage services. |
| Finalize the upgrade | Run the convergence command to refresh your overcloud stack. |
Chapter 2. Preparing for a minor update Copy linkLink copied to clipboard!
You must follow some preparation steps on the undercloud and overcoud before you begin the process to update Red Hat OpenStack Platform 16.0 to the latest minor release.
2.1. Locking the environment to a Red Hat Enterprise Linux release Copy linkLink copied to clipboard!
Red Hat OpenStack Platform 16.0 is supported on Red Hat Enterprise Linux 8.1. Prior to performing the update, lock the undercloud and overcloud repositories to the Red Hat Enterprise Linux 8.1 release to avoid upgrading the operating system to a newer minor release.
Procedure
-
Log into the undercloud as the
stackuser. Source the
stackrcfile:source ~/stackrc
$ source ~/stackrcCopy to Clipboard Copied! Toggle word wrap Toggle overflow Create a static inventory file of your overcloud:
tripleo-ansible-inventory --ansible_ssh_user heat-admin --static-yaml-inventory ~/inventory.yaml
$ tripleo-ansible-inventory --ansible_ssh_user heat-admin --static-yaml-inventory ~/inventory.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow If you use an overcloud name different to the default overcloud name of
overcloud, set the name of your overcloud with the--planoption.Create a playbook that contains a task to lock the operating system version to Red Hat Enterprise Linux 8.1 on all nodes:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow Run the
set_release.yamlplaybook:ansible-playbook -i ~/inventory.yaml -f 25 ~/set_release.yaml
$ ansible-playbook -i ~/inventory.yaml -f 25 ~/set_release.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow
To manually lock a node to a version, log in to the node and run the subscription-manager release command:
sudo subscription-manager release --set=8.1
$ sudo subscription-manager release --set=8.1
2.2. Changing to Extended Update Support (EUS) repositories Copy linkLink copied to clipboard!
Your Red Hat OpenStack Platform subscription includes repositories for Red Hat Enterprise Linux 8.1 Extended Update Support (EUS). The EUS repositories include the latest security patches and bug fixes for Red Hat Enterprise Linux 8.1. Switch to the following repositories before performing a minor version update.
| Standard Repository | EUS Resporitory |
|---|---|
| rhel-8-for-x86_64-baseos-rpms | rhel-8-for-x86_64-baseos-eus-rpms |
| rhel-8-for-x86_64-appstream-rpms | rhel-8-for-x86_64-appstream-eus-rpms |
| rhel-8-for-x86_64-highavailability-rpms | rhel-8-for-x86_64-highavailability-eus-rpms |
Procedure
-
Log into the undercloud as the
stackuser. Source the
stackrcfile:source ~/stackrc
$ source ~/stackrcCopy to Clipboard Copied! Toggle word wrap Toggle overflow -
Edit your overcloud subscription management environment file, which is the file that contains the
RhsmVarsparameter. The default name for this file is usuallyrhsm.yml. Check the
rhsm_reposparameter in your subscription management configuration. If this parameter does not include the EUS repositories, change the relevant repositories to the EUS versions:Copy to Clipboard Copied! Toggle word wrap Toggle overflow - Save the overcloud subscription management environment file.
Create a static inventory file of your overcloud:
tripleo-ansible-inventory --ansible_ssh_user heat-admin --static-yaml-inventory ~/inventory.yaml
$ tripleo-ansible-inventory --ansible_ssh_user heat-admin --static-yaml-inventory ~/inventory.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow If you use an overcloud name different to the default overcloud name of
overcloud, set the name of your overcloud with the--planoption.Create a playbook that contains a task to set the repositories to Red Hat Enterprise Linux 8.1 EUS on all nodes:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow Run the
change_eus.yamlplaybook:ansible-playbook -i ~/inventory.yaml -f 25 ~/change_eus.yaml
$ ansible-playbook -i ~/inventory.yaml -f 25 ~/change_eus.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow
Chapter 3. Updating the Undercloud Copy linkLink copied to clipboard!
This process updates the undercloud and its overcloud images to the latest Red Hat OpenStack Platform 16.0 version.
3.1. Performing a minor update of a containerized undercloud Copy linkLink copied to clipboard!
The director provides commands to update the packages on the undercloud node. This allows you to perform a minor update within the current version of your OpenStack Platform environment.
Procedure
-
Log into the director as the
stackuser. Run
dnfto upgrade the director’s main packages:sudo dnf update -y python3-tripleoclient* openstack-tripleo-common openstack-tripleo-heat-templates tripleo-ansible
$ sudo dnf update -y python3-tripleoclient* openstack-tripleo-common openstack-tripleo-heat-templates tripleo-ansibleCopy to Clipboard Copied! Toggle word wrap Toggle overflow The director uses the
openstack undercloud upgradecommand to update the undercloud environment. Run the command:openstack undercloud upgrade
$ openstack undercloud upgradeCopy to Clipboard Copied! Toggle word wrap Toggle overflow - Wait until the undercloud upgrade process completes.
Reboot the undercloud to update the operating system’s kernel and other system packages:
sudo reboot
$ sudo rebootCopy to Clipboard Copied! Toggle word wrap Toggle overflow - Wait until the node boots.
3.2. Updating the overcloud images Copy linkLink copied to clipboard!
You need to replace your current overcloud images with new versions. The new images ensure the director can introspect and provision your nodes using the latest version of OpenStack Platform software.
Prerequisites
- You have updated the undercloud to the latest version.
Procedure
Source the
stackrcfile:source ~/stackrc
$ source ~/stackrcCopy to Clipboard Copied! Toggle word wrap Toggle overflow Remove any existing images from the
imagesdirectory on thestackuser’s home (/home/stack/images):rm -rf ~/images/*
$ rm -rf ~/images/*Copy to Clipboard Copied! Toggle word wrap Toggle overflow Extract the archives:
cd ~/images for i in /usr/share/rhosp-director-images/overcloud-full-latest-16.0.tar /usr/share/rhosp-director-images/ironic-python-agent-latest-16.0.tar; do tar -xvf $i; done cd ~
$ cd ~/images $ for i in /usr/share/rhosp-director-images/overcloud-full-latest-16.0.tar /usr/share/rhosp-director-images/ironic-python-agent-latest-16.0.tar; do tar -xvf $i; done $ cd ~Copy to Clipboard Copied! Toggle word wrap Toggle overflow Import the latest images into the director:
openstack overcloud image upload --update-existing --image-path /home/stack/images/
$ openstack overcloud image upload --update-existing --image-path /home/stack/images/Copy to Clipboard Copied! Toggle word wrap Toggle overflow Configure your nodes to use the new images:
openstack overcloud node configure $(openstack baremetal node list -c UUID -f value)
$ openstack overcloud node configure $(openstack baremetal node list -c UUID -f value)Copy to Clipboard Copied! Toggle word wrap Toggle overflow Verify the existence of the new images:
openstack image list ls -l /var/lib/ironic/httpboot
$ openstack image list $ ls -l /var/lib/ironic/httpbootCopy to Clipboard Copied! Toggle word wrap Toggle overflow
When deploying overcloud nodes, ensure the overcloud image version corresponds to the respective heat template version. For example, only use the OpenStack Platform 16 images with the OpenStack Platform 16 heat templates.
The new overcloud-full image replaces the old overcloud-full image. If you made changes to the old image, you must repeat the changes in the new image, especially if you want to deploy new nodes in the future.
3.3. Undercloud Post-Upgrade Notes Copy linkLink copied to clipboard!
-
If using a local set of core templates in your
stackusers home directory, ensure you update the templates using the recommended workflow in Using Customized Core Heat Templates in the Advanced Overcloud Customization guide. You must update the local copy before upgrading the overcloud.
3.4. Next Steps Copy linkLink copied to clipboard!
The undercloud upgrade is complete. You can now update the overcloud.
Chapter 4. Updating the Overcloud Copy linkLink copied to clipboard!
This process updates the overcloud.
Prerequisites
- You have updated the undercloud to the latest version.
4.1. Running the overcloud update preparation Copy linkLink copied to clipboard!
The update requires running openstack overcloud update prepare command, which performs the following tasks:
- Updates the overcloud plan to OpenStack Platform 16.0
- Prepares the nodes for the update
Procedure
Source the
stackrcfile:source ~/stackrc
$ source ~/stackrcCopy to Clipboard Copied! Toggle word wrap Toggle overflow Run the update preparation command:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow Include the following options relevant to your environment:
-
Custom configuration environment files (
-e) -
If using your own custom roles, include your custom roles (
roles_data) file (-r) -
If using custom networks, include your composable network (
network_data) file (-n) -
If the name of your overcloud stack is different to the default name
overcloud, include the--stackoption in the update preparation command and replace<STACK_NAME>with the name of your stack.
-
Custom configuration environment files (
- Wait until the update preparation completes.
4.2. Running the container image preparation Copy linkLink copied to clipboard!
The overcloud requires the latest OpenStack Platform 16.0 container images before performing the update. This involves executing the container_image_prepare external update process. To execute this process, run the openstack overcloud external-update run command against tasks tagged with the container_image_prepare tag. These tasks:
- Automatically prepare all container image configuration relevant to your environment.
- Pull the relevant container images to your undercloud, unless you have previously disabled this option.
Procedure
Source the
stackrcfile:source ~/stackrc
$ source ~/stackrcCopy to Clipboard Copied! Toggle word wrap Toggle overflow Run the
openstack overcloud external-update runcommand against tasks tagged with thecontainer_image_preparetag:openstack overcloud external-update run --stack STACK_NAME --tags container_image_prepare
$ openstack overcloud external-update run --stack STACK_NAME --tags container_image_prepareCopy to Clipboard Copied! Toggle word wrap Toggle overflow
4.3. Updating all Controller nodes Copy linkLink copied to clipboard!
This process updates all the Controller nodes to the latest OpenStack Platform 16.0 version. The process involves running the openstack overcloud update run command and including the --limit Controller option to restrict operations to the Controller nodes only.
If you are not using the default stack name (overcloud), set your stack name with the --stack STACK_NAME option replacing STACK_NAME with the name of your stack.
Procedure
Source the
stackrcfile:source ~/stackrc
$ source ~/stackrcCopy to Clipboard Copied! Toggle word wrap Toggle overflow Run the update command:
openstack overcloud update run --stack STACK_NAME --limit Controller --playbook all
$ openstack overcloud update run --stack STACK_NAME --limit Controller --playbook allCopy to Clipboard Copied! Toggle word wrap Toggle overflow - Wait until the Controller node update completes.
4.4. Updating all Compute nodes Copy linkLink copied to clipboard!
This process updates all Compute nodes to the latest OpenStack Platform 16.0 version. The process involves running the openstack overcloud update run command and including the --nodes Compute option to restrict operations to the Compute nodes only.
If you are not using the default stack name (overcloud), set your stack name with the --stack STACK_NAME option replacing STACK_NAME with the name of your stack.
Procedure
Source the
stackrcfile:source ~/stackrc
$ source ~/stackrcCopy to Clipboard Copied! Toggle word wrap Toggle overflow Run the update command:
openstack overcloud update run --stack STACK_NAME --limit Compute --playbook all
$ openstack overcloud update run --stack STACK_NAME --limit Compute --playbook allCopy to Clipboard Copied! Toggle word wrap Toggle overflow - Wait until the Compute node update completes.
4.5. Updating all HCI Compute nodes Copy linkLink copied to clipboard!
This process updates the Hyperconverged Infrastructure (HCI) Compute nodes. The process involves:
-
Running the
openstack overcloud update runcommand and including the--nodes ComputeHCIoption to restrict operations to the HCI nodes only. -
Running the
openstack overcloud external-update run --tags cephcommand to perform an update to a containerized Red Hat Ceph Storage 4 cluster.
If you are not using the default stack name (overcloud), set your stack name with the --stack STACK_NAME option replacing STACK_NAME with the name of your stack.
Procedure
Source the
stackrcfile:source ~/stackrc
$ source ~/stackrcCopy to Clipboard Copied! Toggle word wrap Toggle overflow Run the update command:
openstack overcloud update run --stack _STACK_NAME_ --limit ComputeHCI --playbook all
$ openstack overcloud update run --stack _STACK_NAME_ --limit ComputeHCI --playbook allCopy to Clipboard Copied! Toggle word wrap Toggle overflow - Wait until the node update completes.
Run the Ceph Storage update command. For example:
openstack overcloud external-update run --stack _STACK_NAME_ --tags ceph
$ openstack overcloud external-update run --stack _STACK_NAME_ --tags cephCopy to Clipboard Copied! Toggle word wrap Toggle overflow - Wait until the Compute HCI node update completes.
4.6. Updating all Ceph Storage nodes Copy linkLink copied to clipboard!
This process updates the Ceph Storage nodes. The process involves:
-
Running the
openstack overcloud update runcommand and including the--nodes CephStorageoption to restrict operations to the Ceph Storage nodes only. -
Running the
openstack overcloud external-update runcommand to runceph-ansibleas an external process and update the Red Hat Ceph Storage 4 containers.
If you are not using the default stack name (overcloud), set your stack name with the --stack STACK_NAME option replacing STACK_NAME with the name of your stack.
Procedure
Source the
stackrcfile:source ~/stackrc
$ source ~/stackrcCopy to Clipboard Copied! Toggle word wrap Toggle overflow Update group nodes.
To update all nodes in a group:
openstack overcloud update run --nodes <GROUP_NAME>
$ openstack overcloud update run --nodes <GROUP_NAME>Copy to Clipboard Copied! Toggle word wrap Toggle overflow To update a single node in a group:
openstack overcloud update run --nodes <GROUP_NAME> [NODE_INDEX]
$ openstack overcloud update run --nodes <GROUP_NAME> [NODE_INDEX]Copy to Clipboard Copied! Toggle word wrap Toggle overflow NoteEnsure that you update all nodes if you choose to update nodes individually.
The index of the first node in a group is zero (0). For example, to update the first node in a group named
CephStorage, the command is:openstack overcloud update run --nodes CephStorage[0]- Wait until the node update completes.
Run the Ceph Storage container update command:
openstack overcloud external-update run --stack _STACK_NAME_ --tags ceph
$ openstack overcloud external-update run --stack _STACK_NAME_ --tags cephCopy to Clipboard Copied! Toggle word wrap Toggle overflow - Wait until the Ceph Storage container update completes.
4.7. Performing online database updates Copy linkLink copied to clipboard!
Some overcloud components require an online upgrade (or migration) of their databases tables. This involves executing the online_upgrade external update process. To execute this process, run the openstack overcloud external-update run command against tasks tagged with the online_upgrade tag. This performs online database updates to the following components:
- OpenStack Block Storage (cinder)
- OpenStack Compute (nova)
Procedure
Source the
stackrcfile:source ~/stackrc
$ source ~/stackrcCopy to Clipboard Copied! Toggle word wrap Toggle overflow Run the
openstack overcloud external-update runcommand against tasks tagged with theonline_upgradetag:openstack overcloud external-update run --stack STACK_NAME --tags online_upgrade
$ openstack overcloud external-update run --stack STACK_NAME --tags online_upgradeCopy to Clipboard Copied! Toggle word wrap Toggle overflow
4.8. Finalizing the update Copy linkLink copied to clipboard!
The update requires a final step to update the overcloud stack. This ensures the stack’s resource structure aligns with a regular deployment of OpenStack Platform 16.0 and allows you to perform standard openstack overcloud deploy functions in the future.
Procedure
Source the
stackrcfile:source ~/stackrc
$ source ~/stackrcCopy to Clipboard Copied! Toggle word wrap Toggle overflow Run the update finalization command:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow Include the following options relevant to your environment:
+ * Custom configuration environment files (
-e) * If using your own custom roles, include your custom roles (roles_data) file (-r) * If using custom networks, include your composable network (network_data) file (-n) * If the name of your overcloud stack is different to the default nameovercloud, include the--stackoption in the update preparation command and replace<STACK_NAME>with the name of your stack.- Wait until the update finalization completes.
Chapter 5. Rebooting the overcloud Copy linkLink copied to clipboard!
After a minor Red Hat OpenStack version update, reboot your overcloud. The reboot refreshes the nodes with any associated kernel, system-level, and container component updates. These updates may provide performance and security benefits.
Plan downtime to perform the following reboot procedures.
5.1. Rebooting Controller and composable nodes Copy linkLink copied to clipboard!
Complete the following steps to reboot Controller nodes and standalone nodes based on composable roles, excluding Compute nodes and Ceph Storage nodes.
Procedure
- Log in to the node that you want to reboot.
Optional: If the node uses Pacemaker resources, stop the cluster:
sudo pcs cluster stop
[heat-admin@overcloud-controller-0 ~]$ sudo pcs cluster stopCopy to Clipboard Copied! Toggle word wrap Toggle overflow Reboot the node:
sudo reboot
[heat-admin@overcloud-controller-0 ~]$ sudo rebootCopy to Clipboard Copied! Toggle word wrap Toggle overflow - Wait until the node boots.
Check the services. For example:
If the node uses Pacemaker services, check that the node has rejoined the cluster:
sudo pcs status
[heat-admin@overcloud-controller-0 ~]$ sudo pcs statusCopy to Clipboard Copied! Toggle word wrap Toggle overflow If the node uses Systemd services, check that all services are enabled:
sudo systemctl status
[heat-admin@overcloud-controller-0 ~]$ sudo systemctl statusCopy to Clipboard Copied! Toggle word wrap Toggle overflow If the node uses containerized services, check that all containers on the node are active:
sudo podman ps
[heat-admin@overcloud-controller-0 ~]$ sudo podman psCopy to Clipboard Copied! Toggle word wrap Toggle overflow
5.2. Rebooting a Ceph Storage (OSD) cluster Copy linkLink copied to clipboard!
Complete the following steps to reboot a cluster of Ceph Storage (OSD) nodes.
Procedure
Log in to a Ceph MON or Controller node and disable Ceph Storage cluster rebalancing temporarily:
sudo podman exec -it ceph-mon-controller-0 ceph osd set noout sudo podman exec -it ceph-mon-controller-0 ceph osd set norebalance
$ sudo podman exec -it ceph-mon-controller-0 ceph osd set noout $ sudo podman exec -it ceph-mon-controller-0 ceph osd set norebalanceCopy to Clipboard Copied! Toggle word wrap Toggle overflow - Select the first Ceph Storage node that you want to reboot and log in to the node.
Reboot the node:
sudo reboot
$ sudo rebootCopy to Clipboard Copied! Toggle word wrap Toggle overflow - Wait until the node boots.
Log in to the node and check the cluster status:
sudo podman exec -it ceph-mon-controller-0 ceph status
$ sudo podman exec -it ceph-mon-controller-0 ceph statusCopy to Clipboard Copied! Toggle word wrap Toggle overflow Check that the
pgmapreports allpgsas normal (active+clean).- Log out of the node, reboot the next node, and check its status. Repeat this process until you have rebooted all Ceph storage nodes.
When complete, log into a Ceph MON or Controller node and re-enable cluster rebalancing:
sudo podman exec -it ceph-mon-controller-0 ceph osd unset noout sudo podman exec -it ceph-mon-controller-0 ceph osd unset norebalance
$ sudo podman exec -it ceph-mon-controller-0 ceph osd unset noout $ sudo podman exec -it ceph-mon-controller-0 ceph osd unset norebalanceCopy to Clipboard Copied! Toggle word wrap Toggle overflow Perform a final status check to verify that the cluster reports
HEALTH_OK:sudo podman exec -it ceph-mon-controller-0 ceph status
$ sudo podman exec -it ceph-mon-controller-0 ceph statusCopy to Clipboard Copied! Toggle word wrap Toggle overflow
5.3. Rebooting Compute nodes Copy linkLink copied to clipboard!
Complete the following steps to reboot Compute nodes. To ensure minimal downtime of instances in your Red Hat OpenStack Platform environment, this procedure also includes instructions about migrating instances from the Compute node that you want to reboot. This involves the following workflow:
- Decide whether to migrate instances to another Compute node before rebooting the node.
- Select and disable the Compute node you want to reboot so that it does not provision new instances.
- Migrate the instances to another Compute node.
- Reboot the empty Compute node.
- Enable the empty Compute node.
Prerequisites
Before you reboot the Compute node, you must decide whether to migrate instances to another Compute node while the node is rebooting.
If for some reason you cannot or do not want to migrate the instances, you can set the following core template parameters to control the state of the instances after the Compute node reboots:
NovaResumeGuestsStateOnHostBoot-
Determines whether to return instances to the same state on the Compute node after reboot. When set to
False, the instances will remain down and you must start them manually. Default value is:False NovaResumeGuestsShutdownTimeout-
Number of seconds to wait for an instance to shut down before rebooting. It is not recommended to set this value to
0. Default value is: 300 NovaResumeGuestsShutdownTimeout-
Number of seconds to wait for an instance to shut down before rebooting. It is not recommended to set this value to
0. Default value is: 300
For more information about overcloud parameters and their usage, see Overcloud Parameters.
Procedure
-
Log in to the undercloud as the
stackuser. List all Compute nodes and their UUIDs:
source ~/stackrc (undercloud) $ openstack server list --name compute
$ source ~/stackrc (undercloud) $ openstack server list --name computeCopy to Clipboard Copied! Toggle word wrap Toggle overflow Identify the UUID of the Compute node that you want to reboot.
From the undercloud, select a Compute node. Disable the node:
source ~/overcloudrc (overcloud) $ openstack compute service list (overcloud) $ openstack compute service set [hostname] nova-compute --disable
$ source ~/overcloudrc (overcloud) $ openstack compute service list (overcloud) $ openstack compute service set [hostname] nova-compute --disableCopy to Clipboard Copied! Toggle word wrap Toggle overflow List all instances on the Compute node:
(overcloud) $ openstack server list --host [hostname] --all-projects
(overcloud) $ openstack server list --host [hostname] --all-projectsCopy to Clipboard Copied! Toggle word wrap Toggle overflow - If you decide not to migrate instances, skip to this step.
If you decide to migrate the instances to another Compute node, use one of the following commands:
Migrate the instance to a different host:
(overcloud) $ openstack server migrate [instance-id] --live [target-host]--wait
(overcloud) $ openstack server migrate [instance-id] --live [target-host]--waitCopy to Clipboard Copied! Toggle word wrap Toggle overflow Let
nova-schedulerautomatically select the target host:(overcloud) $ nova live-migration [instance-id]
(overcloud) $ nova live-migration [instance-id]Copy to Clipboard Copied! Toggle word wrap Toggle overflow Live migrate all instances at once:
nova host-evacuate-live [hostname]
$ nova host-evacuate-live [hostname]Copy to Clipboard Copied! Toggle word wrap Toggle overflow NoteThe
novacommand might cause some deprecation warnings, which are safe to ignore.
- Wait until migration completes.
Confirm that the migration was successful:
(overcloud) $ openstack server list --host [hostname] --all-projects
(overcloud) $ openstack server list --host [hostname] --all-projectsCopy to Clipboard Copied! Toggle word wrap Toggle overflow - Continue to migrate instances until none remain on the chosen Compute node.
Log in to the Compute node and reboot the node:
sudo reboot
[heat-admin@overcloud-compute-0 ~]$ sudo rebootCopy to Clipboard Copied! Toggle word wrap Toggle overflow - Wait until the node boots.
Re-enable the Compute node:
source ~/overcloudrc (overcloud) $ openstack compute service set [hostname] nova-compute --enable
$ source ~/overcloudrc (overcloud) $ openstack compute service set [hostname] nova-compute --enableCopy to Clipboard Copied! Toggle word wrap Toggle overflow Check that the Compute node is enabled:
(overcloud) $ openstack compute service list
(overcloud) $ openstack compute service listCopy to Clipboard Copied! Toggle word wrap Toggle overflow
5.4. Rebooting HCI Compute nodes Copy linkLink copied to clipboard!
The following procedure reboots Compute hyperconverged infrastructure (HCI) nodes.
Procedure
Log in to a Ceph MON or Controller node and disable Ceph Storage cluster rebalancing temporarily:
sudo podman exec -it ceph-mon-controller-0 ceph osd set noout sudo podman exec -it ceph-mon-controller-0 ceph osd set norebalance
$ sudo podman exec -it ceph-mon-controller-0 ceph osd set noout $ sudo podman exec -it ceph-mon-controller-0 ceph osd set norebalanceCopy to Clipboard Copied! Toggle word wrap Toggle overflow -
Log in to the undercloud as the
stackuser. List all Compute nodes and their UUIDs:
source ~/stackrc (undercloud) $ openstack server list --name compute
$ source ~/stackrc (undercloud) $ openstack server list --name computeCopy to Clipboard Copied! Toggle word wrap Toggle overflow Identify the UUID of the Compute node you aim to reboot.
From the undercloud, select a Compute node and disable it:
source ~/overcloudrc (overcloud) $ openstack compute service list (overcloud) $ openstack compute service set [hostname] nova-compute --disable
$ source ~/overcloudrc (overcloud) $ openstack compute service list (overcloud) $ openstack compute service set [hostname] nova-compute --disableCopy to Clipboard Copied! Toggle word wrap Toggle overflow List all instances on the Compute node:
(overcloud) $ openstack server list --host [hostname] --all-projects
(overcloud) $ openstack server list --host [hostname] --all-projectsCopy to Clipboard Copied! Toggle word wrap Toggle overflow Use one of the following commands to migrate your instances:
Migrate the instance to a specific host of your choice:
(overcloud) $ openstack server migrate [instance-id] --live [target-host]--wait
(overcloud) $ openstack server migrate [instance-id] --live [target-host]--waitCopy to Clipboard Copied! Toggle word wrap Toggle overflow Let
nova-schedulerautomatically select the target host:(overcloud) $ nova live-migration [instance-id]
(overcloud) $ nova live-migration [instance-id]Copy to Clipboard Copied! Toggle word wrap Toggle overflow Live migrate all instances at once:
nova host-evacuate-live [hostname]
$ nova host-evacuate-live [hostname]Copy to Clipboard Copied! Toggle word wrap Toggle overflow NoteThe
novacommand might cause some deprecation warnings, which are safe to ignore.
- Wait until the migration completes.
Confirm that the migration was successful:
(overcloud) $ openstack server list --host [hostname] --all-projects
(overcloud) $ openstack server list --host [hostname] --all-projectsCopy to Clipboard Copied! Toggle word wrap Toggle overflow - Continue migrating instances until none remain on the chosen Compute node.
Log in to a Ceph MON or a Controller node and check the cluster status:
sudo podman exec $CEPH_MON_CONTAINER ceph -s
$ sudo podman exec $CEPH_MON_CONTAINER ceph -sCopy to Clipboard Copied! Toggle word wrap Toggle overflow Check that the
pgmapreports allpgsas normal (active+clean).Reboot the Compute HCI node:
sudo reboot
$ sudo rebootCopy to Clipboard Copied! Toggle word wrap Toggle overflow - Wait until the node boots.
Enable the Compute node again:
source ~/overcloudrc (overcloud) $ openstack compute service set [hostname] nova-compute --enable
$ source ~/overcloudrc (overcloud) $ openstack compute service set [hostname] nova-compute --enableCopy to Clipboard Copied! Toggle word wrap Toggle overflow Verify that the Compute node is enabled:
(overcloud) $ openstack compute service list
(overcloud) $ openstack compute service listCopy to Clipboard Copied! Toggle word wrap Toggle overflow - Log out of the node, reboot the next node, and check its status. Repeat this process until you have rebooted all Ceph storage nodes.
When complete, log in to a Ceph MON or Controller node and enable cluster rebalancing again:
sudo podman exec $CEPH_MON_CONTAINER ceph osd unset noout sudo podman exec $CEPH_MON_CONTAINER ceph osd unset norebalance
$ sudo podman exec $CEPH_MON_CONTAINER ceph osd unset noout $ sudo podman exec $CEPH_MON_CONTAINER ceph osd unset norebalanceCopy to Clipboard Copied! Toggle word wrap Toggle overflow Perform a final status check to verify the cluster reports
HEALTH_OK:sudo podman exec $CEPH_MON_CONTAINER ceph status
$ sudo podman exec $CEPH_MON_CONTAINER ceph statusCopy to Clipboard Copied! Toggle word wrap Toggle overflow