Working with the model catalog


Red Hat OpenShift AI Cloud Service 1

Working with the model catalog in Red Hat OpenShift AI Cloud Service

Abstract

Discover and evaluate generative AI models in the model catalog and select models to register, deploy, and customize.

Preface

As a data scientist in OpenShift AI, you can discover and evaluate the generative AI models that are available in the model catalog. From the model catalog, you can select the models that you want to register, deploy, and customize.

A model registry acts as a central repository for administrators and data scientists to register, version, and manage the lifecycle of AI models before configuring them for deployment. A model registry is a key component for AI model governance.

The model catalog provides a curated library where data scientists can discover and evaluate the available generative AI models to find the best fit for their use cases.

1.1. Model registry

A model registry is an important component in the lifecycle of an artificial intelligence/machine learning (AI/ML) model, and is a vital part of any machine learning operations (MLOps) platform or workflow. A model registry acts as a central repository, storing metadata related to machine learning models from development to deployment. This metadata ranges from high-level information like the deployment environment and project, to specific details like training hyperparameters, performance metrics, and deployment events.

A model registry acts as a bridge between model experimentation and serving, offering a secure, collaborative metadata store interface for stakeholders in the ML lifecycle. Model registries provide a structured and organized way to store, share, version, deploy, and track models.

OpenShift AI administrators can create model registries in OpenShift AI and grant model registry access to data scientists. For more information, see Managing model registries.

Data scientists with access to a model registry can use it to store, share, version, deploy, and track models. For more information, see Working with model registries.

1.2. Model catalog

Data scientists can use the model catalog to discover and evaluate the models that are available and ready for their organization to register, deploy, and customize.

The model catalog provides models from different providers that data scientists can search and discover before they register models in a model registry and deploy them to a model serving runtime. OpenShift AI administrators can configure the available repository sources for models displayed in the model catalog.

OpenShift AI provides a default model catalog, which includes models from providers such as Red Hat, IBM, Meta, Nvidia, Mistral AI, and Google.

For more information about how data scientists can use the model catalog, see Working with the model catalog.

Chapter 2. Viewing models in the model catalog

You can discover and evaluate the available generative AI models in the model catalog to find the best fit for your use cases.

Prerequisites

  • You are logged in to Red Hat OpenShift AI.

Procedure

  1. From the OpenShift AI dashboard, click ModelsModel catalog.
  2. The Model Catalog page provides a high-level view of available models, including the model name, description, and labels such as task, license, and provider.
  3. In the drop-down list, select from the available catalog sources that have been configured by your administrator. The Default Catalog is displayed by default.

    Note

    OpenShift cluster administrators can configure additional model catalog sources. For more details, see the Kubeflow Model Registry community documentation on configuring catalog sources.

  4. Use the search bar to find a model in the catalog. You can enter text to search by model name, description, or provider.
  5. Click the name of a model to view the model details page. This page displays the model description and the Model card information supplied by the model provider. This includes details such as the model’s intended use and potential limitations, training parameters and datasets, and evaluation results.
  6. You can click Load more models to scroll and view additional models available in the catalog. Repeat this step until all models are loaded.

Verification

  • You can view the information about a selected model on the model details page.

As a data scientist, you can register models directly from the model catalog and create the first version of the new model.

Prerequisites

  • You are logged in to Red Hat OpenShift AI.
  • You have access to an available model registry in your deployment.

Procedure

  1. From the OpenShift AI dashboard, click ModelsModel catalog.
  2. In the drop-down list, select from the available catalog sources that have been configured by your administrator. The Default Catalog is displayed by default.

    Note

    OpenShift cluster administrators can configure additional model catalog sources. For more details, see the Kubeflow Model Registry community documentation on configuring catalog sources.

  3. Use the search bar to find a model in the catalog. You can enter text to search by model name, description, or provider.
  4. Click the name of a model to view the model details page.
  5. Click Register model.
  6. From the Model registry drop-down list, select the model registry that you want to register the model in.
  7. In the Model details section, configure details to apply to all versions of the model:

    1. Optional: In the Model name field, update the name of the model.
    2. Optional: In the Model description field, update the description of the model.
  8. In the Version details section, enter details to apply to the first version of the model:

    1. In the Version name field, enter a name for the model version.
    2. Optional: In the Version description field, enter a description for the first version of the model.
    3. In the Source model format field, enter the name of the model format, for example, ONNX.
    4. In the Source model format version field, enter the version of the model format.
  9. In the Model location section, the URI of the model is displayed.
  10. Click Register model.

Verification

  • The new model details and version are displayed on the Overview tab on the model details page.
  • The new model and version are displayed on the Model registry page.

You can deploy models directly from the model catalog.

Note

OpenShift AI model serving deployments use the global cluster pull secret to pull models in ModelCar format from the catalog.

Prerequisites

Procedure

  1. From the OpenShift AI dashboard, click ModelsModel catalog.
  2. In the drop-down list, select from the available catalog sources that have been configured by your administrator. The Default Catalog is displayed by default.

    Note

    OpenShift cluster administrators can configure additional model catalog sources. For more details, see the Kubeflow Model Registry community documentation on configuring catalog sources.

  3. Use the search bar to find a model in the catalog. You can enter text to search by model name, description, or provider.
  4. Click the name of a model to view the model details page.
  5. Click Deploy model to display the Deploy model dialog.
  6. From the Project drop-down list, select a project in which to deploy your model.

    Note

    Models using OCI storage can only be deployed on the single-model serving platform. Projects using the multi-model serving platform do not appear in the project list.

  7. In the Model deployment section:

    1. Optional: In the Model deployment name field, enter a unique name for your model deployment. This field is autofilled with a value that contains the model name by default.

      This is the name of the inference service created when the model is deployed.

    2. Optional: Click Edit resource name, and then enter a specific resource name for the model deployment in the Resource name field. By default, the resource name matches the name of the model deployment.

      Important

      Resource names are what your resources are labeled as in OpenShift. Your resource name cannot exceed 253 characters, must consist of lowercase alphanumeric characters or -, and must start and end with an alphanumeric character. Resource names are not editable after creation.

      The resource name must not match the name of any other model deployment resource in your OpenShift cluster.

    3. From the Serving runtime list, select a model-serving runtime that is installed and enabled in your OpenShift AI deployment. If project-scoped runtimes exist, the Serving runtime list includes subheadings to distinguish between global runtimes and project-scoped runtimes.
    4. From the Model framework list, select a framework for your model.

      Note

      The Model framework list shows only the frameworks that are supported by the model-serving runtime that you specified when you deployed your model.

  8. From the Deployment mode list, select KServe RawDeployment or Knative Serverless. For more information about deployment modes, see About KServe deployment modes.

    1. In the Number of model server replicas to deploy field, specify a value.
    2. From the Model server size list, select a value.
    3. If you have created a hardware profile, select a hardware profile from the Hardware profile list. If project-scoped hardware profiles exist, the Hardware profile list includes subheadings to distinguish between global hardware profiles and project-scoped hardware profiles.

      Important

      By default, hardware profiles are hidden from appearing in the dashboard navigation menu and user interface. In addition, user interface components associated with the deprecated accelerator profiles functionality are still displayed. To show the Settings → Hardware profiles option in the dashboard navigation menu and the user interface components associated with hardware profiles, set the disableHardwareProfiles value to false in the OdhDashboardConfig custom resource (CR) in OpenShift. For more information about setting dashboard configuration options, see Customizing the dashboard.

    4. In the Model route section, select the Make deployed models available through an external route checkbox to make your deployed models available to external clients.
    5. In the Token authentication section, select the Require token authentication checkbox to require token authentication for your model server. To finish configuring token authentication, perform the following actions:

      1. In the Service account name field, enter a service account name for which the token will be generated. The generated token is created and displayed in the Token secret field when the model server is configured.
      2. To add an additional service account, click Add a service account and enter another service account name.
  9. In the Source model location section, select Current URI to deploy the selected model from the catalog.
  10. Optional: Customize the runtime parameters in the Configuration parameters section:

    1. Modify the values in Additional serving runtime arguments to define how the deployed model behaves.
    2. Modify the values in Additional environment variables to define variables in the model’s environment.
  11. Click Deploy.

Verification

  • The model deployment is displayed on the ModelsModel Deployments page.
  • The model deployment is displayed in the Latest deployments section of the model details page.
  • The model deployment is displayed on the Deployments tab for the model version.

Legal Notice

Copyright © 2025 Red Hat, Inc.
The text of and illustrations in this document are licensed by Red Hat under a Creative Commons Attribution–Share Alike 3.0 Unported license ("CC-BY-SA"). An explanation of CC-BY-SA is available at http://creativecommons.org/licenses/by-sa/3.0/. In accordance with CC-BY-SA, if you distribute this document or an adaptation of it, you must provide the URL for the original version.
Red Hat, as the licensor of this document, waives the right to enforce, and agrees not to assert, Section 4d of CC-BY-SA to the fullest extent permitted by applicable law.
Red Hat, Red Hat Enterprise Linux, the Shadowman logo, the Red Hat logo, JBoss, OpenShift, Fedora, the Infinity logo, and RHCE are trademarks of Red Hat, Inc., registered in the United States and other countries.
Linux® is the registered trademark of Linus Torvalds in the United States and other countries.
Java® is a registered trademark of Oracle and/or its affiliates.
XFS® is a trademark of Silicon Graphics International Corp. or its subsidiaries in the United States and/or other countries.
MySQL® is a registered trademark of MySQL AB in the United States, the European Union and other countries.
Node.js® is an official trademark of Joyent. Red Hat is not formally related to or endorsed by the official Joyent Node.js open source or commercial project.
The OpenStack® Word Mark and OpenStack logo are either registered trademarks/service marks or trademarks/service marks of the OpenStack Foundation, in the United States and other countries and are used with the OpenStack Foundation's permission. We are not affiliated with, endorsed or sponsored by the OpenStack Foundation, or the OpenStack community.
All other trademarks are the property of their respective owners.
Red Hat logoGithubredditYoutubeTwitter

Learn

Try, buy, & sell

Communities

About Red Hat Documentation

We help Red Hat users innovate and achieve their goals with our products and services with content they can trust. Explore our recent updates.

Making open source more inclusive

Red Hat is committed to replacing problematic language in our code, documentation, and web properties. For more details, see the Red Hat Blog.

About Red Hat

We deliver hardened solutions that make it easier for enterprises to work across platforms and environments, from the core datacenter to the network edge.

Theme

© 2026 Red Hat
Back to top