Chapter 3. Deploying OpenShift AI in a disconnected environment
Read this section to understand how to deploy Red Hat OpenShift AI as a development and testing environment for data scientists in a disconnected environment. Disconnected clusters are on a restricted network, typically behind a firewall. In this case, clusters cannot access the remote registries where Red Hat provided OperatorHub sources reside. Instead, the Red Hat OpenShift AI Operator can be deployed to a disconnected environment using a private registry to mirror the images.
Installing OpenShift AI in a disconnected environment involves the following high-level tasks:
- Confirm that your OpenShift cluster meets all requirements. See Requirements for OpenShift AI Self-Managed.
- Add administrative users for OpenShift. See Adding administrative users in OpenShift.
- Mirror images to a private registry. See Mirroring images to a private registry for a disconnected installation.
- Install the Red Hat OpenShift AI Operator. See Installing the Red Hat OpenShift AI Operator.
- Install OpenShift AI components. See Installing and managing Red Hat OpenShift AI components.
- Configure user and administrator groups to provide user access to OpenShift AI. See Adding users to OpenShift AI user groups.
- Provide your users with the URL for the OpenShift cluster on which you deployed OpenShift AI. See Accessing the OpenShift AI dashboard.
- Optionally, configure and enable your accelerators in OpenShift AI to ensure that your data scientists can use compute-heavy workloads in their models. See Enabling accelerators.
3.1. Requirements for OpenShift AI Self-Managed
You must meet the following requirements before you can install Red Hat OpenShift AI on your Red Hat OpenShift cluster in a disconnected environment:
Product subscriptions
You must have a subscription for Red Hat OpenShift AI Self-Managed.
Contact your Red Hat account manager to purchase new subscriptions. If you do not yet have an account manager, complete the form at https://www.redhat.com/en/contact to request one.
Cluster administrator access to your OpenShift cluster
- You must have an OpenShift cluster with cluster administrator access. Use an existing cluster or create a cluster by following the OpenShift Container Platform documentation: Installing a cluster in a disconnected environment.
- After you install a cluster, configure the Cluster Samples Operator by following the OpenShift Container Platform documentation: Configuring Samples Operator for a restricted cluster.
- Your cluster must have at least 2 worker nodes with at least 8 CPUs and 32 GiB RAM available for OpenShift AI to use when you install the Operator. To ensure that OpenShift AI is usable, additional cluster resources are required beyond the minimum requirements.
- To use OpenShift AI on single node OpenShift, the node has to have at least 32 CPUs and 128 GiB RAM.
Your cluster is configured with a default storage class that can be dynamically provisioned.
Confirm that a default storage class is configured by running the
oc get storageclass
command. If no storage classes are noted with(default)
beside the name, follow the OpenShift Container Platform documentation to configure a default storage class: Changing the default storage class. For more information about dynamic provisioning, see Dynamic provisioning.- Open Data Hub must not be installed on the cluster.
For more information about managing the machines that make up an OpenShift cluster, see Overview of machine management.
An identity provider configured for OpenShift
- Red Hat OpenShift AI uses the same authentication systems as Red Hat OpenShift Container Platform. See Understanding identity provider configuration for more information on configuring identity providers.
-
Access to the cluster as a user with the
cluster-admin
role; thekubeadmin
user is not allowed.
Internet access on the mirroring machine
Along with Internet access, the following domains must be accessible to mirror images required for the OpenShift AI Self-Managed installation:
-
cdn.redhat.com
-
subscription.rhn.redhat.com
-
registry.access.redhat.com
-
registry.redhat.io
-
quay.io
-
For CUDA-based images, the following domains must be accessible:
-
ngc.download.nvidia.cn
-
developer.download.nvidia.com
-
Create custom namespaces
-
By default, OpenShift AI uses predefined namespaces, but you can define a custom namespace for the operator and
DSCI.applicationNamespace
as needed. Namespaces created by OpenShift AI typically includeopenshift
orredhat
in their name. Do not rename these system namespaces because they are required for OpenShift AI to function properly. If you are using custom namespaces, before installing the OpenShift AI Operator, you must have created and labeled them as required.
Data science pipelines preparation
- Data science pipelines 2.0 contains an installation of Argo Workflows. If there is an existing installation of Argo Workflows that is not installed by data science pipelines on your cluster, data science pipelines will be disabled after you install OpenShift AI. Before installing OpenShift AI, ensure that your cluster does not have an existing installation of Argo Workflows that is not installed by data science pipelines, or remove the separate installation of Argo Workflows from your cluster.
- Before you can execute a pipeline in a disconnected environment, you must upload the images to your private registry. For more information, see Mirroring images to run pipelines in a restricted environment.
- You can store your pipeline artifacts in an S3-compatible object storage bucket so that you do not consume local storage. To do this, you must first configure write access to your S3 bucket on your storage account.
Install KServe dependencies
- To support the KServe component, which is used by the single-model serving platform to serve large models, you must also install Operators for Red Hat OpenShift Serverless and Red Hat OpenShift Service Mesh and perform additional configuration. For more information, see About the single-model serving platform.
-
If you want to add an authorization provider for the single-model serving platform, you must install the
Red Hat - Authorino
Operator. For information, see Adding an authorization provider for the single-model serving platform.
Install model registry dependencies (Technology Preview feature)
- To use the model registry component, you must also install Operators for Red Hat Authorino, Red Hat OpenShift Serverless, and Red Hat OpenShift Service Mesh. For more information about configuring the model registry component, see Configuring the model registry component.
Access to object storage
- Components of OpenShift AI require or can use S3-compatible object storage such as AWS S3, MinIO, Ceph, or IBM Cloud Storage. An object store is a data storage mechanism that enables users to access their data either as an object or as a file. The S3 API is the recognized standard for HTTP-based access to object storage services.
- The object storage must be accessible to your OpenShift cluster. Deploy the object storage on the same disconnected network as your cluster.
Object storage is required for the following components:
- Single- or multi-model serving platforms, to deploy stored models. See Deploying models on the single-model serving platform or Deploying a model by using the multi-model serving platform.
- Data science pipelines, to store artifacts, logs, and intermediate results. See Configuring a pipeline server and About pipeline logs.
Object storage can be used by the following components:
- Workbenches, to access large datasets. See Adding a connection to your data science project.
- Distributed workloads, to pull input data from and push results to. See Running distributed data science workloads from data science pipelines.
- Code executed inside a pipeline. For example, to store the resulting model in object storage. See Overview of pipelines in Jupyterlab.
3.2. Adding administrative users in OpenShift
Before you can install and configure OpenShift AI for your data scientist users, you must obtain OpenShift cluster administrator (cluster-admin
) privileges.
To assign cluster-admin
privileges to a user, follow the steps in the relevant OpenShift documentation:
- OpenShift Container Platform: Creating a cluster admin
- OpenShift Dedicated: Managing OpenShift Dedicated administrators
- ROSA: Creating a cluster administrator user for quick cluster access
3.3. Mirroring images to a private registry for a disconnected installation
You can install the Red Hat OpenShift AI Operator to your OpenShift cluster in a disconnected environment by mirroring the required container images to a private container registry. After mirroring the images to a container registry, you can install Red Hat OpenShift AI Operator by using OperatorHub.
You can use the mirror registry for Red Hat OpenShift, a small-scale container registry, as a target for mirroring the required container images for OpenShift AI in a disconnected environment. Using the mirror registry for Red Hat OpenShift is optional if another container registry is already available in your installation environment.
Prerequisites
- You have cluster administrator access to a running OpenShift Container Platform cluster, version 4.14 or greater.
- You have credentials for Red Hat OpenShift Cluster Manager (https://console.redhat.com/openshift/).
- Your mirroring machine is running Linux, has 100 GB of space available, and has access to the Internet so that it can obtain the images to populate the mirror repository.
-
You have installed the OpenShift CLI (
oc
). - If you plan to use NVIDIA GPUs, you have mirrored and deployed the NVIDIA GPU Operator. See Configuring the NVIDIA GPU Operator in the OpenShift Container Platform documentation.
- If you plan to use data science pipelines, you have mirrored the OpenShift Pipelines operator.
- If you plan to use the single-model serving platform to serve large models, you have mirrored the Operators for Red Hat OpenShift Serverless and Red Hat OpenShift Service Mesh. For more information, see Serving large models.
- If you plan to use the distributed workloads component, you have mirrored the Ray cluster image.
This procedure uses the oc-mirror plugin v2 (the oc-mirror plugin v1 is now deprecated). For more information, see Changes from oc-mirror plugin v1 to v2 in the OpenShift documentation.
Procedure
- Create a mirror registry. See Creating a mirror registry with mirror registry for Red Hat OpenShift in the OpenShift Container Platform documentation.
To mirror registry images, install the
oc-mirror
OpenShift CLI plugin v2 on your mirroring machine running Linux. See Installing the oc-mirror OpenShift CLI plugin in the OpenShift Container Platform documentation.ImportantThe oc-mirror plugin v1 is deprecated. Red Hat recommends that you use the oc-mirror plugin v2 for continued support and improvements.
- Create a container image registry credentials file that allows mirroring images from Red Hat to your mirror. See Configuring credentials that allow images to be mirrored in the OpenShift Container Platform documentation.
-
Open the example image set configuration file (
rhoai-<version>.md
) from the disconnected installer helper repository and examine its contents. Using the example image set configuration file, create a file called
imageset-config.yaml
and populate it with values suitable for the image set configuration in your deployment.To view a list of the available OpenShift versions, run the following command. This might take several minutes. If the command returns errors, repeat the steps in Configuring credentials that allow images to be mirrored.
Copy to Clipboard Copied! Toggle word wrap Toggle overflow oc-mirror list operators
oc-mirror list operators
To see the available channels for a package in a specific version of OpenShift Container Platform (for example, 4.18), run the following command:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow oc-mirror list operators --catalog=registry.redhat.io/redhat/redhat-operator-index:v4.18 --package=<package_name>
oc-mirror list operators --catalog=registry.redhat.io/redhat/redhat-operator-index:v4.18 --package=<package_name>
For information about subscription update channels, see Understanding update channels.
ImportantThe example image set configurations are for demonstration purposes only and might need further alterations depending on your deployment.
To identify the attributes most suitable for your deployment, examine the documentation and use cases in Mirroring images for a disconnected installation by using the oc-mirror plugin v2.
Your
imageset-config.yaml
should look similar to the following example, whereopenshift-pipelines-operator-rh
is required for data science pipelines, and bothserverless-operator
andservicemeshoperator
are required for the KServe component.Copy to Clipboard Copied! Toggle word wrap Toggle overflow kind: ImageSetConfiguration apiVersion: mirror.openshift.io/v2alpha1 mirror: operators: - catalog: registry.redhat.io/redhat/redhat-operator-index:v4.18 packages: - name: rhods-operator defaultChannel: fast channels: - name: fast minVersion: 2.18.0 maxVersion: 2.18.0 - name: openshift-pipelines-operator-rh channels: - name: stable - name: serverless-operator channels: - name: stable - name: servicemeshoperator channels: - name: stable
kind: ImageSetConfiguration apiVersion: mirror.openshift.io/v2alpha1 mirror: operators: - catalog: registry.redhat.io/redhat/redhat-operator-index:v4.18 packages: - name: rhods-operator defaultChannel: fast channels: - name: fast minVersion: 2.18.0 maxVersion: 2.18.0 - name: openshift-pipelines-operator-rh channels: - name: stable - name: serverless-operator channels: - name: stable - name: servicemeshoperator channels: - name: stable
Download the specified image set configuration to a local file on your mirroring machine:
-
Replace
<mirror_rhoai>
with the target directory where you want to output the image set file. -
The target directory path must start with
file://
. The download might take several minutes.
Copy to Clipboard Copied! Toggle word wrap Toggle overflow oc mirror -c imageset-config.yaml file://<mirror_rhoai> --v2
$ oc mirror -c imageset-config.yaml file://<mirror_rhoai> --v2
TipIf the
tls: failed to verify certificate: x509: certificate signed by unknown authority
error is returned and you want to ignore it, setskipTLS
totrue
in your image set configuration file and run the command again.
-
Replace
Verify that the image set
.tar
files were created:Copy to Clipboard Copied! Toggle word wrap Toggle overflow ls <mirror_rhoai>
$ ls <mirror_rhoai>
Example output
Copy to Clipboard Copied! Toggle word wrap Toggle overflow mirror_000001.tar, mirror_000002.tar
mirror_000001.tar, mirror_000002.tar
If an
archiveSize
value was specified in the image set configuration file, the image set might be separated into multiple.tar
files.Optional: Verify that total size of the image set
.tar
files is around 75 GB:Copy to Clipboard Copied! Toggle word wrap Toggle overflow du -h --max-depth=1 ./<mirror_rhoai>/
$ du -h --max-depth=1 ./<mirror_rhoai>/
If the total size of the image set is significantly less than 75 GB, run the
oc mirror
command again.Upload the contents of the generated image set to your target mirror registry:
-
Replace
<mirror_rhoai>
with the directory that contains your image set.tar
files. Replace
<registry.example.com:5000>
with your mirror registry.Copy to Clipboard Copied! Toggle word wrap Toggle overflow oc mirror -c imageset-config.yaml --from file://<mirror_rhoai> docker://<registry.example.com:5000> --v2
$ oc mirror -c imageset-config.yaml --from file://<mirror_rhoai> docker://<registry.example.com:5000> --v2
TipIf the
tls: failed to verify certificate: x509: certificate signed by unknown authority
error is returned and you want to ignore it, run the following command:Copy to Clipboard Copied! Toggle word wrap Toggle overflow oc mirror --dest-tls-verify false --from=./<mirror_rhoai> docker://<registry.example.com:5000> --v2
$ oc mirror --dest-tls-verify false --from=./<mirror_rhoai> docker://<registry.example.com:5000> --v2
-
Replace
-
Log in to your target OpenShift cluster using the OpenShift CLI as a user with the
cluster-admin
role. Verify that the YAML files are present for the
ImageDigestMirrorSet
andCatalogSource
resources:Replace
<mirror_rhoai>
with the directory that contains your image set.tar
files.Copy to Clipboard Copied! Toggle word wrap Toggle overflow ls <mirror_rhoai>/working-dir/cluster-resources/
$ ls <mirror_rhoai>/working-dir/cluster-resources/
Example output
Copy to Clipboard Copied! Toggle word wrap Toggle overflow cs-redhat-operator-index.yaml idms-oc-mirror.yaml
cs-redhat-operator-index.yaml idms-oc-mirror.yaml
Install the generated resources into the cluster:
Replace
<oc_mirror_workspace_path>
with the path to your oc mirror workspace.Copy to Clipboard Copied! Toggle word wrap Toggle overflow oc apply -f <oc_mirror_workspace_path>/working-dir/cluster-resources
$ oc apply -f <oc_mirror_workspace_path>/working-dir/cluster-resources
Verification
Verify that the
CatalogSource
and pod were created successfully:Copy to Clipboard Copied! Toggle word wrap Toggle overflow oc get catalogsource,pod -n openshift-marketplace
$ oc get catalogsource,pod -n openshift-marketplace
This should return at least one catalog and two pods.
Check that the Red Hat OpenShift AI Operator exists in the OperatorHub:
- Log in to the OpenShift web console.
Click Operators
OperatorHub. The OperatorHub page opens.
- Confirm that the Red Hat OpenShift AI Operator is shown.
- If you mirrored additional operators, such as OpenShift Pipelines, Red Hat OpenShift Serverless, or Red Hat OpenShift Service Mesh, check that those operators exist the OperatorHub.
Additional resources
Mirroring images for a disconnected installation by using the oc-mirror plugin v2
3.4. Configuring custom namespaces
By default, OpenShift AI uses predefined namespaces, but you can define a custom namespace for the operator and DSCI.applicationNamespace
as needed. Namespaces created by OpenShift AI typically include openshift
or redhat
in their name. Do not rename these system namespaces because they are required for OpenShift AI to function properly.
Prerequisites
- You have access to a OpenShift AI cluster with cluster administrator privileges.
- You have downloaded and installed the OpenShift command-line interface (CLI). See Installing the OpenShift CLI.
Procedure
In a terminal window, if you are not already logged in to your OpenShift cluster as a cluster administrator, log in to the OpenShift CLI as shown in the following example:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow oc login <openshift_cluster_url> -u <admin_username> -p <password>
oc login <openshift_cluster_url> -u <admin_username> -p <password>
Enter the following command to create the custom namespace:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow oc create namespace <custom_namespace>
oc create namespace <custom_namespace>
If you are creating a namespace for a
DSCI.applicationNamespace
, enter the following command to add the correct label:Copy to Clipboard Copied! Toggle word wrap Toggle overflow oc label namespace <application_namespace> opendatahub.io/application-namespace=true
oc label namespace <application_namespace> opendatahub.io/application-namespace=true
3.5. Installing the Red Hat OpenShift AI Operator
This section shows how to install the Red Hat OpenShift AI Operator on your OpenShift cluster using the command-line interface (CLI) and the OpenShift web console.
If you want to upgrade from a previous version of OpenShift AI rather than performing a new installation, see Upgrading OpenShift AI in a disconnected environment.
If your OpenShift cluster uses a proxy to access the Internet, you can configure the proxy settings for the Red Hat OpenShift AI Operator. See Overriding proxy settings of an Operator for more information.
3.5.1. Installing the Red Hat OpenShift AI Operator by using the CLI
The following procedure shows how to use the OpenShift command-line interface (CLI) to install the Red Hat OpenShift AI Operator on your OpenShift cluster. You must install the Operator before you can install OpenShift AI components on the cluster.
Prerequisites
- You have a running OpenShift cluster, version 4.14 or greater, configured with a default storage class that can be dynamically provisioned.
- You have cluster administrator privileges for your OpenShift cluster.
- You have downloaded and installed the OpenShift command-line interface (CLI). See Installing the OpenShift CLI.
- You have mirrored the required container images to a private registry. See Mirroring images to a private registry for a disconnected installation.
Procedure
- Open a new terminal window.
Follow these steps to log in to your OpenShift cluster as a cluster administrator:
- In the upper-right corner of the OpenShift web console, click your user name and select Copy login command.
- After you have logged in, click Display token.
Copy the Log in with this token command and paste it in the OpenShift command-line interface (CLI).
Copy to Clipboard Copied! Toggle word wrap Toggle overflow oc login --token=<token> --server=<openshift_cluster_url>
$ oc login --token=<token> --server=<openshift_cluster_url>
Create a namespace for installation of the Operator by performing the following actions:
Create a namespace YAML file named
rhods-operator-namespace.yaml
.Copy to Clipboard Copied! Toggle word wrap Toggle overflow apiVersion: v1 kind: Namespace metadata: name: redhat-ods-operator
apiVersion: v1 kind: Namespace metadata: name: redhat-ods-operator
1 - 1
- Defines the required
redhat-ods-operator
namespace for installation of the Operator.
Create the namespace in your OpenShift cluster.
Copy to Clipboard Copied! Toggle word wrap Toggle overflow oc create -f rhods-operator-namespace.yaml
$ oc create -f rhods-operator-namespace.yaml
You see output similar to the following:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow namespace/redhat-ods-operator created
namespace/redhat-ods-operator created
Create an operator group for installation of the Operator by performing the following actions:
Create an
OperatorGroup
object custom resource (CR) file, for example,rhods-operator-group.yaml
.Copy to Clipboard Copied! Toggle word wrap Toggle overflow apiVersion: operators.coreos.com/v1 kind: OperatorGroup metadata: name: rhods-operator namespace: redhat-ods-operator
apiVersion: operators.coreos.com/v1 kind: OperatorGroup metadata: name: rhods-operator namespace: redhat-ods-operator
1 - 1
- Defines the required
redhat-ods-operator
namespace.
Create the
OperatorGroup
object in your OpenShift cluster.Copy to Clipboard Copied! Toggle word wrap Toggle overflow oc create -f rhods-operator-group.yaml
$ oc create -f rhods-operator-group.yaml
You see output similar to the following:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow operatorgroup.operators.coreos.com/rhods-operator created
operatorgroup.operators.coreos.com/rhods-operator created
Create a subscription for installation of the Operator by performing the following actions:
Create a
Subscription
object CR file, for example,rhods-operator-subscription.yaml
.Copy to Clipboard Copied! Toggle word wrap Toggle overflow apiVersion: operators.coreos.com/v1alpha1 kind: Subscription metadata: name: rhods-operator namespace: redhat-ods-operator spec: name: rhods-operator channel: <channel> source: cs-redhat-operator-index sourceNamespace: openshift-marketplace startingCSV: rhods-operator.x.y.z
apiVersion: operators.coreos.com/v1alpha1 kind: Subscription metadata: name: rhods-operator namespace: redhat-ods-operator
1 spec: name: rhods-operator channel: <channel>
2 source: cs-redhat-operator-index sourceNamespace: openshift-marketplace startingCSV: rhods-operator.x.y.z
3 - 1
- Defines the required
redhat-ods-operator
namespace. - 2
- Sets the update channel. You must specify a value of
fast
,stable
,stable-x.y
eus-x.y
, oralpha
. For more information, see Understanding update channels. - 3
- Optional: Sets the operator version. If you do not specify a value, the subscription defaults to the latest operator version. For more information, see the Red Hat OpenShift AI Self-Managed Life Cycle Knowledgebase article.
Create the
Subscription
object in your OpenShift cluster to install the Operator.Copy to Clipboard Copied! Toggle word wrap Toggle overflow oc create -f rhods-operator-subscription.yaml
$ oc create -f rhods-operator-subscription.yaml
You see output similar to the following:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow subscription.operators.coreos.com/rhods-operator created
subscription.operators.coreos.com/rhods-operator created
Verification
In the OpenShift web console, click Operators
Installed Operators and confirm that the Red Hat OpenShift AI Operator shows one of the following statuses: -
Installing
- installation is in progress; wait for this to change toSucceeded
. This might take several minutes. -
Succeeded
- installation is successful.
-
In the web console, click Home
Projects and confirm that the following project namespaces are visible and listed as Active
:-
redhat-ods-applications
-
redhat-ods-monitoring
-
redhat-ods-operator
-
3.5.2. Installing the Red Hat OpenShift AI Operator by using the web console
The following procedure shows how to use the OpenShift web console to install the Red Hat OpenShift AI Operator on your cluster. You must install the Operator before you can install OpenShift AI components on the cluster.
Prerequisites
- You have a running OpenShift cluster, version 4.14 or greater, configured with a default storage class that can be dynamically provisioned.
- You have cluster administrator privileges for your OpenShift cluster.
- You have mirrored the required container images to a private registry. See Mirroring images to a private registry for a disconnected installation.
Procedure
- Log in to the OpenShift web console as a cluster administrator.
-
In the web console, click Operators
OperatorHub. - On the OperatorHub page, locate the Red Hat OpenShift AI Operator by scrolling through the available Operators or by typing Red Hat OpenShift AI into the Filter by keyword box.
- Click the Red Hat OpenShift AI tile. The Red Hat OpenShift AI information pane opens.
- Select a Channel. For information about subscription update channels, see Understanding update channels.
- Select a Version.
- Click Install. The Install Operator page opens.
- Review or change the selected channel and version as needed.
-
For Installation mode, note that the only available value is
All namespaces on the cluster (default)
. This installation mode makes the Operator available to all namespaces in the cluster. -
For Installed Namespace, select
Operator recommended Namespace: redhat-ods-operator
. For Update approval, select one of the following update strategies:
-
Automatic
: Your environment attempts to install new updates when they are available based on the content of your mirror. Manual
: A cluster administrator must approve any new updates before installation begins.ImportantBy default, the Red Hat OpenShift AI Operator follows a sequential update process. This means that if there are several versions between the current version and the target version, Operator Lifecycle Manager (OLM) upgrades the Operator to each of the intermediate versions before it upgrades it to the final, target version.
If you configure automatic upgrades, OLM automatically upgrades the Operator to the latest available version. If you configure manual upgrades, a cluster administrator must manually approve each sequential update between the current version and the final, target version.
For information about supported versions, see the Red Hat OpenShift AI Life Cycle Knowledgebase article.
-
Click Install.
The Installing Operators pane appears. When the installation finishes, a checkmark appears next to the Operator name.
Verification
In the OpenShift web console, click Operators
Installed Operators and confirm that the Red Hat OpenShift AI Operator shows one of the following statuses: -
Installing
- installation is in progress; wait for this to change toSucceeded
. This might take several minutes. -
Succeeded
- installation is successful.
-
In the web console, click Home
Projects and confirm that the following project namespaces are visible and listed as Active
:-
redhat-ods-applications
-
redhat-ods-monitoring
-
redhat-ods-operator
-
3.6. Installing and managing Red Hat OpenShift AI components
You can use the OpenShift command-line interface (CLI) or OpenShift web console to install and manage components of Red Hat OpenShift AI on your OpenShift cluster.
3.6.1. Installing Red Hat OpenShift AI components by using the CLI
To install Red Hat OpenShift AI components by using the OpenShift command-line interface (CLI), you must create and configure a DataScienceCluster
object.
The following procedure describes how to create and configure a DataScienceCluster
object to install Red Hat OpenShift AI components as part of a new installation.
- For information about changing the installation status of OpenShift AI components after installation, see Updating the installation status of Red Hat OpenShift AI components by using the web console.
- For information about upgrading OpenShift AI, see Upgrading OpenShift AI Self-Managed in a disconnected environment.
Prerequisites
- The Red Hat OpenShift AI Operator is installed on your OpenShift cluster. See Installing the Red Hat OpenShift AI Operator.
- You have cluster administrator privileges for your OpenShift cluster.
- You have downloaded and installed the OpenShift command-line interface (CLI). See Installing the OpenShift CLI.
Procedure
- Open a new terminal window.
Follow these steps to log in to your OpenShift cluster as a cluster administrator:
- In the upper-right corner of the OpenShift web console, click your user name and select Copy login command.
- After you have logged in, click Display token.
Copy the Log in with this token command and paste it in the OpenShift command-line interface (CLI).
Copy to Clipboard Copied! Toggle word wrap Toggle overflow oc login --token=<token> --server=<openshift_cluster_url>
$ oc login --token=<token> --server=<openshift_cluster_url>
Create a
DataScienceCluster
object custom resource (CR) file, for example,rhods-operator-dsc.yaml
.Copy to Clipboard Copied! Toggle word wrap Toggle overflow apiVersion: datasciencecluster.opendatahub.io/v1 kind: DataScienceCluster metadata: name: default-dsc spec: components: codeflare: managementState: Removed dashboard: managementState: Removed datasciencepipelines: managementState: Removed kserve: managementState: Removed kueue: managementState: Removed modelmeshserving: managementState: Removed ray: managementState: Removed trainingoperator: managementState: Removed trustyai: managementState: Removed workbenches: managementState: Removed
apiVersion: datasciencecluster.opendatahub.io/v1 kind: DataScienceCluster metadata: name: default-dsc spec: components: codeflare: managementState: Removed dashboard: managementState: Removed datasciencepipelines: managementState: Removed kserve: managementState: Removed
1 2 kueue: managementState: Removed modelmeshserving: managementState: Removed ray: managementState: Removed trainingoperator: managementState: Removed trustyai: managementState: Removed workbenches: managementState: Removed
- 1
- To fully install the KServe component, which is used by the single-model serving platform to serve large models, you must install Operators for Red Hat OpenShift Service Mesh and Red Hat OpenShift Serverless and perform additional configuration. See Installing the single-model serving platform.
- 2
- If you have not enabled the KServe component (that is, you set the value of the
managementState
field toRemoved
), you must also disable the dependent Service Mesh component to avoid errors. See Disabling KServe dependencies.
In the
spec.components
section of the CR, for each OpenShift AI component shown, set the value of themanagementState
field to eitherManaged
orRemoved
. These values are defined as follows:- Managed
- The Operator actively manages the component, installs it, and tries to keep it active. The Operator will upgrade the component only if it is safe to do so.
- Removed
- The Operator actively manages the component but does not install it. If the component is already installed, the Operator will try to remove it.
Important- To learn how to fully install the KServe component, which is used by the single-model serving platform to serve large models, see Installing the single-model serving platform.
-
If you have not enabled the KServe component (that is, you set the value of the
managementState
field toRemoved
), you must also disable the dependent Service Mesh component to avoid errors. See Disabling KServe dependencies. - To learn how to install the distributed workloads components, see Installing the distributed workloads components.
- To learn how to run distributed workloads in a disconnected environment, see Running distributed data science workloads in a disconnected environment.
Create the
DataScienceCluster
object in your OpenShift cluster to install the specified OpenShift AI components.Copy to Clipboard Copied! Toggle word wrap Toggle overflow oc create -f rhods-operator-dsc.yaml
$ oc create -f rhods-operator-dsc.yaml
You see output similar to the following:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow datasciencecluster.datasciencecluster.opendatahub.io/default created
datasciencecluster.datasciencecluster.opendatahub.io/default created
Verification
Confirm that there is a running pod for each component:
-
In the OpenShift web console, click Workloads
Pods. -
In the Project list at the top of the page, select
redhat-ods-applications
. - In the applications namespace, confirm that there are running pods for each of the OpenShift AI components that you installed.
-
In the OpenShift web console, click Workloads
Confirm the status of all installed components:
-
In the OpenShift web console, click Operators
Installed Operators. - Click the Red Hat OpenShift AI Operator.
-
Click the Data Science Cluster tab and select the
DataScienceCluster
object calleddefault-dsc
. - Select the YAML tab.
In the
installedComponents
section, confirm that the components you installed have a status value oftrue
.NoteIf a component shows with the
component-name: {}
format in thespec.components
section of the CR, the component is not installed.
-
In the OpenShift web console, click Operators
- In the Red Hat OpenShift AI dashboard, users can view the list of the installed OpenShift AI components, their corresponding source (upstream) components, and the versions of the installed components, as described in Viewing installed components.
3.6.2. Installing Red Hat OpenShift AI components by using the web console
To install Red Hat OpenShift AI components by using the OpenShift web console, you must create and configure a DataScienceCluster
object.
The following procedure describes how to create and configure a DataScienceCluster
object to install Red Hat OpenShift AI components as part of a new installation.
- For information about changing the installation status of OpenShift AI components after installation, see Updating the installation status of Red Hat OpenShift AI components by using the web console.
- For information about upgrading OpenShift AI, see Upgrading OpenShift AI Self-Managed in a disconnected environment.
Prerequisites
- The Red Hat OpenShift AI Operator is installed on your OpenShift cluster. See Installing the Red Hat OpenShift AI Operator.
- You have cluster administrator privileges for your OpenShift cluster.
Procedure
- Log in to the OpenShift web console as a cluster administrator.
-
In the web console, click Operators
Installed Operators and then click the Red Hat OpenShift AI Operator. - Click the Data Science Cluster tab.
- Click Create DataScienceCluster.
For Configure via, select YAML view.
An embedded YAML editor opens showing a default custom resource (CR) for the
DataScienceCluster
object, similar to the following example:Copy to Clipboard Copied! Toggle word wrap Toggle overflow apiVersion: datasciencecluster.opendatahub.io/v1 kind: DataScienceCluster metadata: name: default-dsc spec: components: codeflare: managementState: Removed dashboard: managementState: Removed datasciencepipelines: managementState: Removed kserve: managementState: Removed kueue: managementState: Removed modelmeshserving: managementState: Removed ray: managementState: Removed trainingoperator: managementState: Removed trustyai: managementState: Removed workbenches: managementState: Removed
apiVersion: datasciencecluster.opendatahub.io/v1 kind: DataScienceCluster metadata: name: default-dsc spec: components: codeflare: managementState: Removed dashboard: managementState: Removed datasciencepipelines: managementState: Removed kserve: managementState: Removed
1 2 kueue: managementState: Removed modelmeshserving: managementState: Removed ray: managementState: Removed trainingoperator: managementState: Removed trustyai: managementState: Removed workbenches: managementState: Removed
- 1
- To fully install the KServe component, which is used by the single-model serving platform to serve large models, you must install Operators for Red Hat OpenShift Service Mesh and Red Hat OpenShift Serverless and perform additional configuration. See Installing the single-model serving platform.
- 2
- If you have not enabled the KServe component (that is, you set the value of the
managementState
field toRemoved
), you must also disable the dependent Service Mesh component to avoid errors. See Disabling KServe dependencies.
In the
spec.components
section of the CR, for each OpenShift AI component shown, set the value of themanagementState
field to eitherManaged
orRemoved
. These values are defined as follows:- Managed
- The Operator actively manages the component, installs it, and tries to keep it active. The Operator will upgrade the component only if it is safe to do so.
- Removed
- The Operator actively manages the component but does not install it. If the component is already installed, the Operator will try to remove it.
Important- To learn how to fully install the KServe component, which is used by the single-model serving platform to serve large models, see Installing the single-model serving platform.
-
If you have not enabled the KServe component (that is, you set the value of the
managementState
field toRemoved
), you must also disable the dependent Service Mesh component to avoid errors. See Disabling KServe dependencies. - To learn how to install the distributed workloads components, see Installing the distributed workloads components.
- To learn how to run distributed workloads in a disconnected environment, see Running distributed data science workloads in a disconnected environment.
- Click Create.
Verification
Confirm that there is a running pod for each component:
-
In the OpenShift web console, click Workloads
Pods. -
In the Project list at the top of the page, select
redhat-ods-applications
. - In the applications namespace, confirm that there are running pods for each of the OpenShift AI components that you installed.
-
In the OpenShift web console, click Workloads
Confirm the status of all installed components:
-
In the OpenShift web console, click Operators
Installed Operators. - Click the Red Hat OpenShift AI Operator.
-
Click the Data Science Cluster tab and select the
DataScienceCluster
object calleddefault-dsc
. - Select the YAML tab.
In the
installedComponents
section, confirm that the components you installed have a status value oftrue
.NoteIf a component shows with the
component-name: {}
format in thespec.components
section of the CR, the component is not installed.
-
In the OpenShift web console, click Operators
- In the Red Hat OpenShift AI dashboard, users can view the list of the installed OpenShift AI components, their corresponding source (upstream) components, and the versions of the installed components, as described in Viewing installed components.
3.6.3. Updating the installation status of Red Hat OpenShift AI components by using the web console
You can use the OpenShift web console to update the installation status of components of Red Hat OpenShift AI on your OpenShift cluster.
If you upgraded OpenShift AI, the upgrade process automatically used the values of the previous version’s DataScienceCluster
object. New components are not automatically added to the DataScienceCluster
object.
After upgrading OpenShift AI:
-
Inspect the default
DataScienceCluster
object to check and optionally update themanagementState
status of the existing components. -
Add any new components to the
DataScienceCluster
object.
Prerequisites
- The Red Hat OpenShift AI Operator is installed on your OpenShift cluster.
- You have cluster administrator privileges for your OpenShift cluster.
Procedure
- Log in to the OpenShift web console as a cluster administrator.
-
In the web console, click Operators
Installed Operators and then click the Red Hat OpenShift AI Operator. - Click the Data Science Cluster tab.
-
On the DataScienceClusters page, click the
default
object. Click the YAML tab.
An embedded YAML editor opens showing the default custom resource (CR) for the
DataScienceCluster
object, similar to the following example:Copy to Clipboard Copied! Toggle word wrap Toggle overflow apiVersion: datasciencecluster.opendatahub.io/v1 kind: DataScienceCluster metadata: name: default-dsc spec: components: codeflare: managementState: Removed dashboard: managementState: Removed datasciencepipelines: managementState: Removed kserve: managementState: Removed kueue: managementState: Removed modelmeshserving: managementState: Removed ray: managementState: Removed trainingoperator: managementState: Removed trustyai: managementState: Removed workbenches: managementState: Removed
apiVersion: datasciencecluster.opendatahub.io/v1 kind: DataScienceCluster metadata: name: default-dsc spec: components: codeflare: managementState: Removed dashboard: managementState: Removed datasciencepipelines: managementState: Removed kserve: managementState: Removed kueue: managementState: Removed modelmeshserving: managementState: Removed ray: managementState: Removed trainingoperator: managementState: Removed trustyai: managementState: Removed workbenches: managementState: Removed
In the
spec.components
section of the CR, for each OpenShift AI component shown, set the value of themanagementState
field to eitherManaged
orRemoved
. These values are defined as follows:- Managed
- The Operator actively manages the component, installs it, and tries to keep it active. The Operator will upgrade the component only if it is safe to do so.
- Removed
- The Operator actively manages the component but does not install it. If the component is already installed, the Operator will try to remove it.
Important- To learn how to install the KServe component, which is used by the single-model serving platform to serve large models, see Installing the single-model serving platform.
-
If you have not enabled the KServe component (that is, you set the value of the
managementState
field toRemoved
), you must also disable the dependent Service Mesh component to avoid errors. See Disabling KServe dependencies. - To learn how to install the distributed workloads feature, see Installing the distributed workloads components.
- To learn how to run distributed workloads in a disconnected environment, see Running distributed data science workloads in a disconnected environment.
Click Save.
For any components that you updated, OpenShift AI initiates a rollout that affects all pods to use the updated image.
Verification
Confirm that there is a running pod for each component:
-
In the OpenShift web console, click Workloads
Pods. -
In the Project list at the top of the page, select
redhat-ods-applications
. - In the applications namespace, confirm that there are running pods for each of the OpenShift AI components that you installed.
-
In the OpenShift web console, click Workloads
Confirm the status of all installed components:
-
In the OpenShift web console, click Operators
Installed Operators. - Click the Red Hat OpenShift AI Operator.
-
Click the Data Science Cluster tab and select the
DataScienceCluster
object calleddefault-dsc
. - Select the YAML tab.
In the
installedComponents
section, confirm that the components you installed have a status value oftrue
.NoteIf a component shows with the
component-name: {}
format in thespec.components
section of the CR, the component is not installed.
-
In the OpenShift web console, click Operators
- In the Red Hat OpenShift AI dashboard, users can view the list of the installed OpenShift AI components, their corresponding source (upstream) components, and the versions of the installed components, as described in Viewing installed components.
3.6.4. Viewing installed OpenShift AI components
In the Red Hat OpenShift AI dashboard, you can view a list of the installed OpenShift AI components, their corresponding source (upstream) components, and the versions of the installed components.
Prerequisites
- OpenShift AI is installed in your OpenShift cluster.
Procedure
- Log in to the OpenShift AI dashboard.
-
In the top navigation bar, click the help icon (
) and then select About.
Verification
The About page shows a list of the installed OpenShift AI components along with their corresponding upstream components and upstream component versions.
Additional resources