이 콘텐츠는 선택한 언어로 제공되지 않습니다.

Chapter 1. Version 3.3 release notes


Red Hat Enterprise Linux AI is a generative AI inference platform for Linux environments that uses Red Hat AI Inference Server for running and optimizing models, and includes Red Hat AI Model Optimization Toolkit for model quantization, sparsity, and general compression for supported AI accelerators. Red Hat AI Model Optimization Toolkit has native Hugging Face and vLLM support. You can seamlessly integrate optimized models with deployment pipelines for faster, cost-saving inference at scale, powered by the compressed-tensors model format.

Red Hat Enterprise Linux AI is packaged as a bootc container image for easy deployment on a Linux server appliance with NVIDIA CUDA or AMD ROCm AI accelerators installed. The container images are available from registry.redhat.io:

  • registry.redhat.io/rhelai3/bootc-cuda-rhel9:3.3.0
  • registry.redhat.io/rhelai3/bootc-rocm-rhel9:3.3.0
Important

There is no direct upgrade path from Red Hat Enterprise Linux AI 1.5 to Red Hat Enterprise Linux AI 3.0. You can upgrade from Red Hat Enterprise Linux AI 3.0 to 3.3 and all versions in-between.

Important

The registry.redhat.io/rhelai3/bootc-rocm-rhel9:3.3.0 image does not include Red Hat AI Model Optimization Toolkit, which is not supported for AMD ROCm AI accelerators.

1.1. New features

Red Hat Enterprise Linux AI 3.3 packages Red Hat AI Inference Server 3.3, which includes the following highlights:

New model support
Red Hat AI Inference Server 3.3 adds support for Mistral 3 models including Mixture of Experts (MoE) architecture variants, IBM Prithvi geospatial foundation models, and various other models including BAGEL, AudioFlamingo3, and JAIS 2.
New AI accelerator support
Red Hat AI Inference Server 3.3 adds support for NVIDIA B300 and GB300 Blackwell AI accelerators with CUDA 13.0, AMD Instinct MI325X AI accelerators, and CPU-only x86_64 AVX2 inference as a Technology Preview. Support for AWS Trainium and Inferentia accelerators is also available as a Technology Preview.
Performance improvements
Whisper models now run approximately 3 times faster compared to the previous release. DeepSeek-V3.1 models provide 5.3% throughput improvement and 4.4% time-to-first-token improvement.
Model optimization updates
Red Hat AI Model Optimization Toolkit adds model-free post-training quantization on safetensors files, extended KV cache and attention quantization capabilities, and the AutoRoundModifier algorithm.

For the complete list of new features, enhancements, and known issues, see the Red Hat AI Inference Server 3.3 release notes.

1.2. Known issues

There are no known issues for Red Hat Enterprise Linux AI 3.3.

Red Hat logoGithubredditYoutubeTwitter

자세한 정보

평가판, 구매 및 판매

커뮤니티

Red Hat 문서 정보

Red Hat을 사용하는 고객은 신뢰할 수 있는 콘텐츠가 포함된 제품과 서비스를 통해 혁신하고 목표를 달성할 수 있습니다. 최신 업데이트를 확인하세요.

보다 포괄적 수용을 위한 오픈 소스 용어 교체

Red Hat은 코드, 문서, 웹 속성에서 문제가 있는 언어를 교체하기 위해 최선을 다하고 있습니다. 자세한 내용은 다음을 참조하세요.Red Hat 블로그.

Red Hat 소개

Red Hat은 기업이 핵심 데이터 센터에서 네트워크 에지에 이르기까지 플랫폼과 환경 전반에서 더 쉽게 작업할 수 있도록 강화된 솔루션을 제공합니다.

Theme

© 2026 Red Hat
맨 위로 이동