Chapter 7. Supported deployment environments
The following deployment environments for Red Hat AI Inference Server are supported.
Red Hat AI Inference Server is available only as a container image. The host operating system and kernel must support the required accelerator drivers. For more information, see Supported AI accelerators.
| Environment | Supported versions | Deployment notes |
|---|---|---|
| OpenShift Container Platform (self‑managed) | 4.14+ | Deploy on bare‑metal hosts or virtual machines. |
| Red Hat OpenShift Service on AWS (ROSA) | 4.14+ | Requires a ROSA cluster with STS and GPU‑enabled P5 or G5 node types. See Prepare your environment for more information. |
| Red Hat Enterprise Linux AI | 3.0+ | Deploy on bare‑metal hosts or virtual machines. |
| Red Hat Enterprise Linux (RHEL) | 9.2+ | Deploy on bare‑metal hosts or virtual machines. |
| Linux (not RHEL) | - | Supported under third‑party policy deployed on bare‑metal hosts or virtual machines. OpenShift Container Platform Operators are not required. |
| Kubernetes (not OpenShift Container Platform) | - | Supported under third‑party policy deployed on bare‑metal hosts or virtual machines. |
- Single-host deployments for IBM Spyre AI accelerators on IBM Z and IBM Power are supported for RHEL AI 9.6+.
- Cluster deployments for IBM Spyre AI accelerators on IBM Z are supported as part of Red Hat OpenShift AI version 3.0+ only.