Questo contenuto non è disponibile nella lingua selezionata.

Chapter 1. Version 3.3 release notes


Red Hat Enterprise Linux AI is a generative AI inference platform for Linux environments that uses Red Hat AI Inference Server for running and optimizing models, and includes Red Hat AI Model Optimization Toolkit for model quantization, sparsity, and general compression for supported AI accelerators. Red Hat AI Model Optimization Toolkit has native Hugging Face and vLLM support. You can seamlessly integrate optimized models with deployment pipelines for faster, cost-saving inference at scale, powered by the compressed-tensors model format.

Red Hat Enterprise Linux AI is packaged as a bootc container image for easy deployment on a Linux server appliance with NVIDIA CUDA or AMD ROCm AI accelerators installed. The container images are available from registry.redhat.io:

  • registry.redhat.io/rhelai3/bootc-cuda-rhel9:3.3.0
  • registry.redhat.io/rhelai3/bootc-rocm-rhel9:3.3.0
Important

There is no direct upgrade path from Red Hat Enterprise Linux AI 1.5 to Red Hat Enterprise Linux AI 3.0. You can upgrade from Red Hat Enterprise Linux AI 3.0 to 3.3 and all versions in-between.

Important

The registry.redhat.io/rhelai3/bootc-rocm-rhel9:3.3.0 image does not include Red Hat AI Model Optimization Toolkit, which is not supported for AMD ROCm AI accelerators.

1.1. New features

Red Hat Enterprise Linux AI 3.3 packages Red Hat AI Inference Server 3.3, which includes the following highlights:

New model support
Red Hat AI Inference Server 3.3 adds support for Mistral 3 models including Mixture of Experts (MoE) architecture variants, IBM Prithvi geospatial foundation models, and various other models including BAGEL, AudioFlamingo3, and JAIS 2.
New AI accelerator support
Red Hat AI Inference Server 3.3 adds support for NVIDIA B300 and GB300 Blackwell AI accelerators with CUDA 13.0, AMD Instinct MI325X AI accelerators, and CPU-only x86_64 AVX2 inference as a Technology Preview. Support for AWS Trainium and Inferentia accelerators is also available as a Technology Preview.
Performance improvements
Whisper models now run approximately 3 times faster compared to the previous release. DeepSeek-V3.1 models provide 5.3% throughput improvement and 4.4% time-to-first-token improvement.
Model optimization updates
Red Hat AI Model Optimization Toolkit adds model-free post-training quantization on safetensors files, extended KV cache and attention quantization capabilities, and the AutoRoundModifier algorithm.

For the complete list of new features, enhancements, and known issues, see the Red Hat AI Inference Server 3.3 release notes.

1.2. Known issues

There are no known issues for Red Hat Enterprise Linux AI 3.3.

Red Hat logoGithubredditYoutubeTwitter

Formazione

Prova, acquista e vendi

Community

Informazioni sulla documentazione di Red Hat

Aiutiamo gli utenti Red Hat a innovarsi e raggiungere i propri obiettivi con i nostri prodotti e servizi grazie a contenuti di cui possono fidarsi. Esplora i nostri ultimi aggiornamenti.

Rendiamo l’open source più inclusivo

Red Hat si impegna a sostituire il linguaggio problematico nel codice, nella documentazione e nelle proprietà web. Per maggiori dettagli, visita il Blog di Red Hat.

Informazioni su Red Hat

Forniamo soluzioni consolidate che rendono più semplice per le aziende lavorare su piattaforme e ambienti diversi, dal datacenter centrale all'edge della rete.

Theme

© 2026 Red Hat
Torna in cima