Chapter 3. Scaling storage capacity of AWS OpenShift Data Foundation cluster
3.1. Scaling up storage capacity of AWS OpenShift Data Foundation cluster Copy linkLink copied to clipboard!
To increase the storage capacity in a dynamically created storage cluster on an user-provisioned infrastructure, you can add storage capacity and performance to your configured Red Hat OpenShift Data Foundation worker nodes.
You can scale up storage capacity of an AWS Red Hat OpenShift Data Foundation cluster in two ways:
3.1.1. Scaling up storage capacity on a cluster by adding a new set of OSDs Copy linkLink copied to clipboard!
To scale the storage capacity of your configured Red Hat OpenShift Data Foundation worker nodes, you can increase the capacity by adding three disks at a time. Three disks are needed since OpenShift Data Foundation uses a replica count of 3 to maintain the high availability. So the amount of storage consumed is three times the usable space.
Usable space might vary when encryption is enabled or replica 2 pools are being used. To increase the storage capacity in a dynamically created storage cluster on an user-provisioned infrastructure, you can add storage capacity and performance to your configured Red Hat OpenShift Data Foundation worker nodes.
Prerequisites
- You have administrative privilege to the OpenShift Container Platform Console.
- You have a running OpenShift Data Foundation Storage Cluster.
- The disk should be of the same size and type as used during initial deployment.
Procedure
- Log in to the OpenShift Web Console.
-
Click Operators
Installed Operators. - Click OpenShift Data Foundation Operator.
Click the Storage Systems tab.
- Click the Action Menu (⋮) on the far right of the storage system name to extend the options menu.
- Select Add Capacity from the options menu.
- Select the Storage Class. Choose the storage class which you wish to use to provision new storage devices.
- Click Add.
-
To check the status, navigate to Storage
Data Foundation and verify that the Storage System
in the Status card has a green tick.
Verification steps
Verify the Raw Capacity card.
-
In the OpenShift Web Console, click Storage
Data Foundation. - In the Status card of the Overview tab, click Storage System and then click the storage system link from the pop up that appears.
In the Block and File tab, check the Raw Capacity card.
Note that the capacity increases based on your selections.
NoteThe raw capacity does not take replication into account and shows the full capacity.
-
In the OpenShift Web Console, click Storage
Verify that the new object storage devices (OSDs) and their corresponding new Persistent Volume Claims (PVCs) are created.
To view the state of the newly created OSDs:
-
Click Workloads
Pods from the OpenShift Web Console. Select
openshift-storage
from the Project drop-down list.NoteIf the Show default projects option is disabled, use the toggle button to list all the default projects.
-
Click Workloads
To view the state of the PVCs:
-
Click Storage
Persistent Volume Claims from the OpenShift Web Console. Select
openshift-storage
from the Project drop-down list.NoteIf the Show default projects option is disabled, use the toggle button to list all the default projects.
-
Click Storage
Optional: If cluster-wide encryption is enabled on the cluster, verify that the new OSD devices are encrypted.
Identify the nodes where the new OSD pods are running.
oc get -n openshift-storage -o=custom-columns=NODE:.spec.nodeName pod/<OSD-pod-name>
$ oc get -n openshift-storage -o=custom-columns=NODE:.spec.nodeName pod/<OSD-pod-name>
Copy to Clipboard Copied! Toggle word wrap Toggle overflow <OSD-pod-name>
Is the name of the OSD pod.
For example:
oc get -n openshift-storage -o=custom-columns=NODE:.spec.nodeName pod/rook-ceph-osd-0-544db49d7f-qrgqm
$ oc get -n openshift-storage -o=custom-columns=NODE:.spec.nodeName pod/rook-ceph-osd-0-544db49d7f-qrgqm
Copy to Clipboard Copied! Toggle word wrap Toggle overflow Example output:
NODE compute-1
NODE compute-1
Copy to Clipboard Copied! Toggle word wrap Toggle overflow
For each of the nodes identified in the previous step, do the following:
Create a debug pod and open a chroot environment for the selected hosts.
oc debug node/<node-name>
$ oc debug node/<node-name>
Copy to Clipboard Copied! Toggle word wrap Toggle overflow <node-name>
Is the name of the node.
chroot /host
$ chroot /host
Copy to Clipboard Copied! Toggle word wrap Toggle overflow
Check for the
crypt
keyword beside theocs-deviceset
names.lsblk
$ lsblk
Copy to Clipboard Copied! Toggle word wrap Toggle overflow
Cluster reduction is supported only with the Red Hat Support Team’s assistance.
3.1.2. Scaling up storage capacity on an AWS cluster by resizing existing OSDs Copy linkLink copied to clipboard!
To increase the storage capacity on a cluster, you can add storage capacity by resizing existing OSDs.
Resizing existing OSDs can only be done once every 6 hours due to AWS limitations. If you exceed this amount within the 6 hour time frame, you will receive a warning that includes the note You’ve reached the maximum modification rate per volume limit. Wait at least 6 hours between modifications per EBS volume.
Prerequisites
- You have administrative privilege to the OpenShift Container Platform Console.
- You have a running OpenShift Data Foundation Storage Cluster.
Procedure
Update the
dataPVCTemplate
size for thestorageDeviceSets
with the new desired size using theoc patch
command.Copy to Clipboard Copied! Toggle word wrap Toggle overflow In this example YAML, the
storage
parameter understorageDeviceSets
reflects the current size of512Gi
.Using the
oc patch
command:Get the current OSD storage for the
storageDeviceSets
you are increasing storage for:oc get storagecluster ocs-storagecluster -n openshift-storage -o jsonpath=' {.spec.storageDeviceSets[0].dataPVCTemplate.spec.resources.requests.storage} ' 512Gi
oc get storagecluster ocs-storagecluster -n openshift-storage -o jsonpath=' {.spec.storageDeviceSets[0].dataPVCTemplate.spec.resources.requests.storage} ' 512Gi
Copy to Clipboard Copied! Toggle word wrap Toggle overflow Increase the storage with the desired value (the following example reflect the size change of 2Ti):
Copy to Clipboard Copied! Toggle word wrap Toggle overflow
- Wait for the OSDs to restart.
Confirm that the resize took effect:
oc get pvc -l ceph.rook.io/DeviceSet -n openshift-storage
$ oc get pvc -l ceph.rook.io/DeviceSet -n openshift-storage
Copy to Clipboard Copied! Toggle word wrap Toggle overflow Verify that for all the resized OSDs, resize is completed and reflected correctly in the
CAPACITY
column of the command output.- If the resize did not take effect, restart the OSD pods again. It may take multiple restarts for the resize to complete.
3.2. Scaling out storage capacity on a AWS cluster Copy linkLink copied to clipboard!
OpenShift Data Foundation is highly scalable. It can be scaled out by adding new nodes with required storage and enough hardware resources in terms of CPU and RAM. Practically there is no limit on the number of nodes which can be added but from the support perspective 2000 nodes is the limit for OpenShift Data Foundation.
Scaling out storage capacity can be broken down into two steps
- Adding new node
- Scaling up the storage capacity
OpenShift Data Foundation does not support heterogeneous OSD/Disk sizes.
3.2.1. Adding a node Copy linkLink copied to clipboard!
You can add nodes to increase the storage capacity when existing worker nodes are already running at their maximum supported OSDs or there are not enough resources to add new OSDs on the existing nodes. It is always recommended to add nodes in the multiple of three, each of them in different failure domains.
While it is recommended to add nodes in the multiple of three, you still have the flexibility to add one node at a time in the flexible scaling deployment. Refer to the Knowledgebase article Verify if flexible scaling is enabled.
OpenShift Data Foundation does not support heterogeneous disk size and types. The new nodes to be added should have the disk of the same type and size which was used during OpenShift Data Foundation deployment.
3.2.1.1. Adding a node to an installer-provisioned infrastructure Copy linkLink copied to clipboard!
Prerequisites
- You have administrative privilege to the OpenShift Container Platform Console.
- You have a running OpenShift Data Foundation Storage Cluster.
Procedure
-
Navigate to Compute
Machine Sets. On the machine set where you want to add nodes, select Edit Machine Count.
- Add the amount of nodes, and click Save.
-
Click Compute
Nodes and confirm if the new node is in Ready state.
Apply the OpenShift Data Foundation label to the new node.
-
For the new node, click Action menu (⋮)
Edit Labels. - Add cluster.ocs.openshift.io/openshift-storage, and click Save.
-
For the new node, click Action menu (⋮)
It is recommended to add 3 nodes, one each in different zones. You must add 3 nodes and perform this procedure for all of them. In case of bare metal installer-provisioned infrastructure deployment, you must expand the cluster first. For instructions, see Expanding the cluster.
Verification steps
Execute the following command in the terminal and verify that the new node is present in the output:
oc get nodes --show-labels | grep cluster.ocs.openshift.io/openshift-storage= |cut -d' ' -f1
$ oc get nodes --show-labels | grep cluster.ocs.openshift.io/openshift-storage= |cut -d' ' -f1
Copy to Clipboard Copied! Toggle word wrap Toggle overflow On the OpenShift web console, click Workloads
Pods, confirm that at least the following pods on the new node are in Running state: -
csi-cephfsplugin-*
-
csi-rbdplugin-*
-
To scale up storage capacity:
- For dynamic storage devices, see Scaling up storage capacity on a cluster.
3.2.1.2. Adding a node to an user-provisioned infrastructure Copy linkLink copied to clipboard!
Prerequisites
- You have administrative privilege to the OpenShift Container Platform Console.
- You have a running OpenShift Data Foundation Storage Cluster.
Procedure
Depending on the type of infrastructure, perform the following steps:
- Get a new machine with the required infrastructure. See Platform requirements.
- Create a new OpenShift Container Platform worker node using the new machine.
Check for certificate signing requests (CSRs) that are in
Pending
state.oc get csr
$ oc get csr
Copy to Clipboard Copied! Toggle word wrap Toggle overflow Approve all the required CSRs for the new node.
oc adm certificate approve <Certificate_Name>
$ oc adm certificate approve <Certificate_Name>
Copy to Clipboard Copied! Toggle word wrap Toggle overflow <Certificate_Name>
- Is the name of the CSR.
-
Click Compute
Nodes, confirm if the new node is in Ready state. Apply the OpenShift Data Foundation label to the new node using any one of the following:
- From User interface
-
For the new node, click Action Menu (⋮)
Edit Labels. -
Add
cluster.ocs.openshift.io/openshift-storage
, and click Save.
-
For the new node, click Action Menu (⋮)
- From Command line interface
Apply the OpenShift Data Foundation label to the new node.
oc label node <new_node_name> cluster.ocs.openshift.io/openshift-storage=""
$ oc label node <new_node_name> cluster.ocs.openshift.io/openshift-storage=""
Copy to Clipboard Copied! Toggle word wrap Toggle overflow <new_node_name>
- Is the name of the new node.
Verification steps
Execute the following command in the terminal and verify that the new node is present in the output:
oc get nodes --show-labels | grep cluster.ocs.openshift.io/openshift-storage= |cut -d' ' -f1
$ oc get nodes --show-labels | grep cluster.ocs.openshift.io/openshift-storage= |cut -d' ' -f1
Copy to Clipboard Copied! Toggle word wrap Toggle overflow On the OpenShift web console, click Workloads
Pods, confirm that at least the following pods on the new node are in Running state: -
csi-cephfsplugin-*
-
csi-rbdplugin-*
-
To scale up storage capacity:
- For dynamic storage devices, see Scaling up storage capacity on a cluster.
3.2.2. Scaling up storage capacity Copy linkLink copied to clipboard!
To scale up storage capacity, see Scaling up storage capacity on a cluster.