Chapter 6. Working with hardware profiles
Hardware profiles are currently available in Red Hat OpenShift AI as a Technology Preview feature only. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process.
For more information about the support scope of Red Hat Technology Preview features, see Technology Preview Features Support Scope
In Red Hat OpenShift AI, you can schedule user workloads on worker nodes that have specific hardware configurations, such as hardware accelerators, CPU-only nodes, or specialized memory allocations. With hardware profiles, you can define these hardware resources explicitly, enabling precise targeting of workloads to specific nodes and improving resource management efficiency.
By default, this feature is hidden from appearing in the dashboard navigation menu. To show the Settings disableHardwareProfiles
value to false
in the OdhDashboardConfig
custom resource (CR) in OpenShift. For more information, see Dashboard configuration options.
You can use hardware profiles to create profiles with hardware identifiers, explicit resource allocation limits (CPU, memory, and accelerators), tolerations, and node selectors. These capabilities are particularly beneficial in environments with heterogeneous hardware, including multiple GPU types, CPU-only configurations, memory-intensive workloads, or even single-node deployments. This targeted scheduling significantly enhances resource utilization, reduces overhead, and optimizes costs, especially in complex environments, such as clusters with diverse hardware.
To get started, contact your cluster administrator to identify hardware resources available in your cluster.
To configure specific hardware configurations for your data scientists to use in OpenShift AI, you must create an associated hardware profile. A hardware profile is a custom resource definition (CRD) on OpenShift that has a HardwareProfile resource, and defines the hardware specification. You can create and manage hardware profiles by selecting Settings
After you create and enable a hardware profile, users can select the hardware profile in the user interface when deploying workbenches, model-serving workloads, and pipelines, where applicable.
6.1. Creating a hardware profile
To configure specific hardware configurations for your data scientists to use in OpenShift AI, you must create an associated hardware profile.
Prerequisites
- You have logged in to OpenShift AI as a user with OpenShift AI administrator privileges.
- The relevant hardware is installed and you have confirmed that it is detected in your environment.
Procedure
From the OpenShift AI dashboard, click Settings
Hardware profiles. The Hardware profiles page appears, displaying existing hardware profiles. To enable or disable an existing hardware profile, on the row containing the relevant hardware profile, click the toggle in the Enabled column.
Click Create hardware profile.
The Create hardware profile page appears.
- In the Name field, enter a name for the hardware profile.
- Optional: To change the default name of your Kubernetes resource, click Edit resource name and enter a name in the Resource name field. The resource name cannot be edited after creation.
- Optional: In the Description field, enter a description for the hardware profile.
In the Visiblity section, set the hardware profile visibility level:
- To access the hardware profile in all areas of OpenShift AI, leave the Visible everywhere radio button selected.
- Click the Limited visibility radio button to limit the areas of OpenShift AI where your data scientists can use the hardware profile.
Optional: Configure node resource request limits:
Click Add resource.
The Add resource dialog opens.
- In the Resource label field, enter a unique resource label.
- In the Resource identifier field, enter a unique resource identifier.
- From the Resource type field, select a resource type from the list.
- In the Default field, enter the default resource request limit. This value must be equal to or between the minimum and maximum limits.
- In the Minimum allowed field, enter the minimum number of resources that users can request.
In the Maximum allowed field, enter the maximum number of resources that users can request:
- To set a specific maximum request limit, click the Set maximum limit radio button and enter a value.
- To set no maximum request limit, click the No maximum limit radio button.
- Click Add.
Optional: Add a node selector to schedule pods on nodes with matching labels.
Click Add node selector.
The Add node selector dialog opens.
- In the Key field, enter a node selection key. The key must begin with a letter or number, and may contain letters, numbers, hyphens, dots, and underscores.
- In the Value field, enter a node selection value. The value must begin with a letter or number, and may contain letters, numbers, hyphens, dots, and underscores.
- Click Add.
Optional: Add a toleration to schedule pods with matching taints.
Click Add toleration.
The Add toleration dialog opens.
From the Operator list, select one of the following options:
- Equal - The key/value/effect parameters must match. This is the default.
- Exists - The key/effect parameters must match. You must leave a blank value parameter, which matches any.
From the Effect list, select one of the following options:
- None
- NoSchedule - New pods that do not match the taint are not scheduled onto that node. Existing pods on the node remain.
- PreferNoSchedule - New pods that do not match the taint might be scheduled onto that node, but the scheduler tries not to. Existing pods on the node remain.
- NoExecute - New pods that do not match the taint cannot be scheduled onto that node. Existing pods on the node that do not have a matching toleration are removed.
- In the Key field, enter a toleration key. The key must begin with a letter or number, and may contain letters, numbers, hyphens, dots, and underscores.
- In the Value field, enter a toleration value. The value must begin with a letter or number, and may contain letters, numbers, hyphens, dots, and underscores.
In the Toleration Seconds section, select one of the following options to specify how long a pod stays bound to a node that has a node condition:
- Forever - Pods stays permanently bound to a node.
- Custom value - Enter a value, in seconds, to define how long pods stay bound to a node that has a node condition.
- Click Add.
- Click Create hardware profile.
Verification
- The hardware profile appears on the Hardware profiles page.
- The hardware profile appears in the Hardware profiles list on the Create workbench page.
-
The hardware profile appears on the Instances tab on the details page for the
HardwareProfile
custom resource definition (CRD).
6.2. Updating a hardware profile
You can update the existing hardware profiles in your deployment. You can change important identifying information, such as the display name, the identifier, or the description.
Prerequisites
- You have logged in to OpenShift AI as a user with OpenShift AI administrator privileges.
- The hardware profile exists in your deployment.
Procedure
From the OpenShift AI dashboard, click Settings
Hardware profiles. The Hardware profiles page appears. Existing hardware profiles are displayed. To enable or disable a hardware profile, on the row containing the relevant hardware profile, click the toggle in the Enabled column.
Click the action menu (⋮) and select Edit from the list.
The Edit hardware profile dialog opens.
- Make your changes.
- Click Update hardware profile.
Verification
- If your hardware profile has new identifying information, this information appears in the Hardware profile list on the Create workbench page.
6.3. Deleting a hardware profile
To discard hardware profiles that you no longer require, you can delete them so that they do not appear on the dashboard.
Prerequisites
- You have logged in to OpenShift AI as a user with OpenShift AI administrator privileges.
- The hardware profile that you want to delete exists in your deployment.
Procedure
From the OpenShift AI dashboard, click Settings
Hardware profiles. The Hardware profiles page appears, displaying existing hardware profiles.
Click the action menu (⋮) beside the hardware profile that you want to delete and click Delete.
The Delete hardware profile dialog opens.
- Enter the name of the hardware profile in the text field to confirm that you intend to delete it.
- Click Delete.
Verification
- The hardware profile no longer appears on the Hardware profiles page.
6.4. Configuring a recommended accelerator for workbench images
To help you indicate the most suitable accelerators to your data scientists, you can configure a recommended tag to appear on the dashboard.
Prerequisites
- You have logged in to OpenShift AI as a user with OpenShift AI administrator privileges.
- You have existing workbench images in your deployment.
- You have enabled GPU support in OpenShift AI. This includes installing the Node Feature Discovery operator and NVIDIA GPU Operators. For more information, see Installing the Node Feature Discovery operator and Enabling NVIDIA GPUs.
Procedure
From the OpenShift AI dashboard, click Settings
Workbench images. The Workbench images page appears. Previously imported workbench images are displayed.
Click the action menu (⋮) and select Edit from the list.
The Update workbench image dialog opens.
- From the Accelerator identifier list, select an identifier to set its accelerator as recommended with the workbench image. If the workbench image contains only one accelerator identifier, the identifier name displays by default.
Click Update.
NoteIf you have already configured an accelerator identifier for a workbench image, you can specify a recommended accelerator for the workbench image by creating an associated hardware profile. To do this, click Create profile on the row containing the workbench image and complete the relevant fields. If the workbench image does not contain an accelerator identifier, you must manually configure one before creating an associated hardware profile.
Verification
- When your data scientists select an accelerator with a specific workbench image, a tag appears next to the corresponding accelerator indicating its compatibility.
6.5. Configuring a recommended accelerator for serving runtimes
To help you indicate the most suitable accelerators to your data scientists, you can configure a recommended accelerator tag for your serving runtimes.
Prerequisites
- You have logged in to OpenShift AI as a user with OpenShift AI administrator privileges.
- You have enabled GPU support in OpenShift AI. This includes installing the Node Feature Discovery operator and NVIDIA GPU Operators. For more information, see Installing the Node Feature Discovery operator and Enabling NVIDIA GPUs.
Procedure
From the OpenShift AI dashboard, click Settings
Serving runtimes. The Serving runtimes page opens and shows the model-serving runtimes that are already installed and enabled in your OpenShift AI deployment. By default, the OpenVINO Model Server runtime is pre-installed and enabled in OpenShift AI.
Edit your custom runtime that you want to add the recommended accelerator tag to, click the action menu (⋮) and select Edit.
A page with an embedded YAML editor opens.
NoteYou cannot directly edit the OpenVINO Model Server runtime that is included in OpenShift AI by default. However, you can clone this runtime and edit the cloned version. You can then add the edited clone as a new, custom runtime. To do this, click the action menu beside the OpenVINO Model Server and select Duplicate.
In the editor, enter the YAML code to apply the annotation
opendatahub.io/recommended-accelerators
. The excerpt in this example shows the annotation to set a recommended tag for an NVIDIA GPU accelerator:metadata: annotations: opendatahub.io/recommended-accelerators: '["nvidia.com/gpu"]'
metadata: annotations: opendatahub.io/recommended-accelerators: '["nvidia.com/gpu"]'
Copy to Clipboard Copied! - Click Update.
Verification
- When your data scientists select an accelerator with a specific serving runtime, a tag appears next to the corresponding accelerator indicating its compatibility.