Ce contenu n'est pas disponible dans la langue sélectionnée.

Chapter 1. Version 3.3 release notes


Red Hat Enterprise Linux AI is a generative AI inference platform for Linux environments that uses Red Hat AI Inference Server for running and optimizing models, and includes Red Hat AI Model Optimization Toolkit for model quantization, sparsity, and general compression for supported AI accelerators. Red Hat AI Model Optimization Toolkit has native Hugging Face and vLLM support. You can seamlessly integrate optimized models with deployment pipelines for faster, cost-saving inference at scale, powered by the compressed-tensors model format.

Red Hat Enterprise Linux AI is packaged as a bootc container image for easy deployment on a Linux server appliance with NVIDIA CUDA or AMD ROCm AI accelerators installed. The container images are available from registry.redhat.io:

  • registry.redhat.io/rhelai3/bootc-cuda-rhel9:3.3.0
  • registry.redhat.io/rhelai3/bootc-rocm-rhel9:3.3.0
Important

There is no direct upgrade path from Red Hat Enterprise Linux AI 1.5 to Red Hat Enterprise Linux AI 3.0. You can upgrade from Red Hat Enterprise Linux AI 3.0 to 3.3 and all versions in-between.

Important

The registry.redhat.io/rhelai3/bootc-rocm-rhel9:3.3.0 image does not include Red Hat AI Model Optimization Toolkit, which is not supported for AMD ROCm AI accelerators.

1.1. New features

Red Hat Enterprise Linux AI 3.3 packages Red Hat AI Inference Server 3.3, which includes the following highlights:

New model support
Red Hat AI Inference Server 3.3 adds support for Mistral 3 models including Mixture of Experts (MoE) architecture variants, IBM Prithvi geospatial foundation models, and various other models including BAGEL, AudioFlamingo3, and JAIS 2.
New AI accelerator support
Red Hat AI Inference Server 3.3 adds support for NVIDIA B300 and GB300 Blackwell AI accelerators with CUDA 13.0, AMD Instinct MI325X AI accelerators, and CPU-only x86_64 AVX2 inference as a Technology Preview. Support for AWS Trainium and Inferentia accelerators is also available as a Technology Preview.
Performance improvements
Whisper models now run approximately 3 times faster compared to the previous release. DeepSeek-V3.1 models provide 5.3% throughput improvement and 4.4% time-to-first-token improvement.
Model optimization updates
Red Hat AI Model Optimization Toolkit adds model-free post-training quantization on safetensors files, extended KV cache and attention quantization capabilities, and the AutoRoundModifier algorithm.

For the complete list of new features, enhancements, and known issues, see the Red Hat AI Inference Server 3.3 release notes.

1.2. Known issues

There are no known issues for Red Hat Enterprise Linux AI 3.3.

Red Hat logoGithubredditYoutubeTwitter

Apprendre

Essayez, achetez et vendez

Communautés

À propos de la documentation Red Hat

Nous aidons les utilisateurs de Red Hat à innover et à atteindre leurs objectifs grâce à nos produits et services avec un contenu auquel ils peuvent faire confiance. Découvrez nos récentes mises à jour.

Rendre l’open source plus inclusif

Red Hat s'engage à remplacer le langage problématique dans notre code, notre documentation et nos propriétés Web. Pour plus de détails, consultez le Blog Red Hat.

À propos de Red Hat

Nous proposons des solutions renforcées qui facilitent le travail des entreprises sur plusieurs plates-formes et environnements, du centre de données central à la périphérie du réseau.

Theme

© 2026 Red Hat
Retour au début