Red Hat AI Inference Server 3.2
Related Products
Red Hat Enterprise Linux AI
Find release notes and product documentation for using the Enterprise Linux AI foundation model platform to consistently develop, test, and run large language models (LLMs) in enterprise applications
Red Hat OpenShift AI
Find release notes and product documentation for using the OpenShift AI platform and its integrated MLOps capabilities to manage predictive and generative AI models at scale
Welcome
Release notes
Highlights of what is new and what has changed with this Red Hat AI Inference Server release
Getting started
Getting started with Red Hat AI Inference Server
Validated models
Red Hat AI Inference Server Validated models
Supported product and hardware configurations
Supported hardware and software configurations for deploying Red Hat AI Inference Server
OpenShift Container Platform에 Red Hat AI Inference Server 배포
AI 액셀러레이터가 설치된 OpenShift Container Platform 클러스터에 Red Hat AI Inference Server 배포
Deploying Red Hat AI Inference Server in a disconnected environment
Deploy Red Hat AI Inference Server in a disconnected environment using OpenShift Container Platform and a disconnected mirror image registry
Inference Operations
Inference serving language models in OCI-compliant model containers
Inferencing OCI-compliant models in Red Hat AI Inference Server
vLLM server arguments
Server arguments for running Red Hat AI Inference Server
LLM Compressor
Compressing large language models with the LLM Compressor library
Additional Resources
Product life cycle
Understand the product life cycle to plan deployments and support applications using the product
Red Hat AI learning hub
Explore a curated collection of learning resources designed to help you accomplish key tasks with Red Hat AI products and services