Chapter 3. Supported deployment environments


The following deployment environments for Red Hat AI Inference Server are supported.

Table 3.1. Red Hat AI Inference Server supported deployment environments
EnvironmentSupported versionsDeployment notes

OpenShift Container Platform (self‑managed)

4.14 – 4.18

Deploy on bare‑metal hosts or virtual machines.

Red Hat OpenShift Service on AWS (ROSA)

4.14 – 4.18

Requires ROSA STS cluster with GPU‑enabled P5 or G5 node types.

Red Hat Enterprise Linux (RHEL)

9.2 – 10.0

Deploy on bare‑metal hosts or virtual machines.

Linux (not RHEL)

-

Supported under third‑party policy deployed on bare‑metal hosts or virtual machines. OpenShift Container Platform Operators are not required.

Kubernetes (not OpenShift Container Platform)

-

Supported under third‑party policy deployed on bare‑metal hosts or virtual machines.

Note

Red Hat AI Inference Server is available only as a container image. The host operating system and kernel must support the required accelerator drivers. For more information, see Supported AI accelerators.

Back to top
Red Hat logoGithubredditYoutubeTwitter

Learn

Try, buy, & sell

Communities

About Red Hat Documentation

We help Red Hat users innovate and achieve their goals with our products and services with content they can trust. Explore our recent updates.

Making open source more inclusive

Red Hat is committed to replacing problematic language in our code, documentation, and web properties. For more details, see the Red Hat Blog.

About Red Hat

We deliver hardened solutions that make it easier for enterprises to work across platforms and environments, from the core datacenter to the network edge.

Theme

© 2025 Red Hat