Getting started
Getting started with Red Hat AI Inference Server
Abstract
Chapter 1. About AI Inference Server Copy linkLink copied to clipboard!
AI Inference Server provides enterprise-grade stability and security, building on the open source vLLM project, which provides state-of-the-art inferencing features.
AI Inference Server uses continuous batching to process requests as they arrive instead of waiting for a full batch to be accumulated. It also uses tensor parallelism to distribute LLM workloads across multiple GPUs. These features provide reduced latency and higher throughput.
To reduce the cost of inferencing models, AI Inference Server uses paged attention. LLMs use a mechanism called attention to understand conversations with users. Normally, attention uses a significant amount of memory, much of which is wasted. Paged attention addresses this memory waste by provisioning memory for LLMs similar to the way that virtual memory works for operating systems. This approach consumes less memory and lowers costs.
Red Hat AI Inference Server is available as a container image from the Red Hat container registry. You can browse available images in the Red Hat Ecosystem Catalog.
To find Red Hat AI Inference Server container images in the Red Hat Ecosystem Catalog, search for "AI Inference Server".
To verify cost savings and performance gains with AI Inference Server, complete the following procedures:
- Serving and inferencing with AI Inference Server
- Validating Red Hat AI Inference Server benefits using key metrics
Chapter 2. Product and version compatibility Copy linkLink copied to clipboard!
The following table lists the supported product versions for Red Hat AI Inference Server, Red Hat Enterprise Linux AI, and Red Hat OpenShift AI.
| Product version | vLLM core version | LLM Compressor version |
|---|---|---|
| 3.4.0-ea.1 | v0.14.1 | v0.9.0.2 |
| 3.3 | v0.13.0 | v0.9.0.1 |
| 3.2.5 | v0.11.2 | v0.8.1 |
| 3.2.4 | v0.11.0 | v0.8.1 |
| 3.2.3 | v0.11.0 | v0.8.1 |
| 3.2.2 | v0.10.1.1 | v0.7.1 |
| 3.2.1 | v0.10.0 | Not included in this release |
| 3.2.0 | v0.9.2 | Not included in this release |
| Product version | vLLM core version | LLM Compressor version |
|---|---|---|
| 3.3 | v0.13.0 | v0.9.0.1 |
| 3.2 | v0.11.2 | v0.8.1 |
| 3.0 | v0.11.0 | v0.8.1 |
| Product version | vLLM core version | LLM Compressor version |
|---|---|---|
| 3.3 | v0.13.0 | v0.9.0.1 |
| 3.2 | v0.11.2 | v0.8.1 |
| 3.0 | v0.11.0 | v0.8.1 |
Chapter 3. Reviewing AI Inference Server Python packages Copy linkLink copied to clipboard!
You can review the Python packages installed in the Red Hat AI Inference Server container image by running the container with Podman and reviewing the pip list package output.
Prerequisites
- You have installed Podman or Docker.
- You are logged in as a user with sudo access.
-
You have access to
registry.redhat.ioand have logged in.
Procedure
Run the Red Hat AI Inference Server container image with the
pip list packagecommand to view all installed Python packages. For example:$ podman run --rm --entrypoint=/bin/bash \ registry.redhat.io/rhaii-early-access/vllm-cuda-rhel9:3.4.0-ea.1 \ -c "pip list"To view detailed information about a specific package, run the Podman command with
pip show <package_name>. For example:$ podman run --rm --entrypoint=/bin/bash \ registry.redhat.io/rhaii-early-access/vllm-cuda-rhel9:3.4.0-ea.1 \ -c "pip show vllm"Example output
Name: vllm Version: v0.14.1
Chapter 4. Serving and inferencing with Podman using NVIDIA CUDA AI accelerators Copy linkLink copied to clipboard!
Serve and inference a large language model with Podman and Red Hat AI Inference Server running on NVIDIA CUDA AI accelerators.
Prerequisites
- You have installed Podman or Docker.
- You are logged in as a user with sudo access.
-
You have access to
registry.redhat.ioand have logged in. - You have a Hugging Face account and have generated a Hugging Face access token.
You have access to a Linux server with data center grade NVIDIA AI accelerators installed.
For NVIDIA GPUs:
- Install NVIDIA drivers
- Install the NVIDIA Container Toolkit
- If your system has multiple NVIDIA GPUs that use NVSwitch, you must have root access to start Fabric Manager
For more information about supported vLLM quantization schemes for accelerators, see Supported hardware.
Procedure
Open a terminal on your server host, and log in to
registry.redhat.io:$ podman login registry.redhat.ioPull the relevant the NVIDIA CUDA image by running the following command:
$ podman pull registry.redhat.io/rhaii-early-access/vllm-cuda-rhel9:3.4.0-ea.1If your system has SELinux enabled, configure SELinux to allow device access:
$ sudo setsebool -P container_use_devices 1Create a volume and mount it into the container. Adjust the container permissions so that the container can use it.
$ mkdir -p rhaii-cache$ chmod g+rwX rhaii-cacheCreate or append your
HF_TOKENHugging Face token to theprivate.envfile. Source theprivate.envfile.$ echo "export HF_TOKEN=<your_HF_token>" > private.env$ source private.envStart the AI Inference Server container image.
For NVIDIA CUDA accelerators, if the host system has multiple GPUs and uses NVSwitch, then start NVIDIA Fabric Manager. To detect if your system is using NVSwitch, first check if files are present in
/proc/driver/nvidia-nvswitch/devices/, and then start NVIDIA Fabric Manager. Starting NVIDIA Fabric Manager requires root privileges.$ ls /proc/driver/nvidia-nvswitch/devices/Example output
0000:0c:09.0 0000:0c:0a.0 0000:0c:0b.0 0000:0c:0c.0 0000:0c:0d.0 0000:0c:0e.0$ systemctl start nvidia-fabricmanagerImportantNVIDIA Fabric Manager is only required on systems with multiple GPUs that use NVSwitch. For more information, see NVIDIA Server Architectures.
Check that the Red Hat AI Inference Server container can access NVIDIA GPUs on the host by running the following command:
$ podman run --rm -it \ --security-opt=label=disable \ --device nvidia.com/gpu=all \ nvcr.io/nvidia/cuda:12.4.1-base-ubi9 \ nvidia-smiExample output
+-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 570.124.06 Driver Version: 570.124.06 CUDA Version: 12.8 | |-----------------------------------------+------------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 NVIDIA A100-SXM4-80GB Off | 00000000:08:01.0 Off | 0 | | N/A 32C P0 64W / 400W | 1MiB / 81920MiB | 0% Default | | | | Disabled | +-----------------------------------------+------------------------+----------------------+ | 1 NVIDIA A100-SXM4-80GB Off | 00000000:08:02.0 Off | 0 | | N/A 29C P0 63W / 400W | 1MiB / 81920MiB | 0% Default | | | | Disabled | +-----------------------------------------+------------------------+----------------------+ +-----------------------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=========================================================================================| | No running processes found | +-----------------------------------------------------------------------------------------+Start the container.
$ podman run --rm -it \ --device nvidia.com/gpu=all \ --security-opt=label=disable \ --shm-size=4g -p 8000:8000 \ --userns=keep-id:uid=1001 \ --env "HUGGING_FACE_HUB_TOKEN=$HF_TOKEN" \ --env "HF_HUB_OFFLINE=0" \ -v ./rhaii-cache:/opt/app-root/src/.cache:Z \ registry.redhat.io/rhaii-early-access/vllm-cuda-rhel9:3.4.0-ea.1 \ --model RedHatAI/Llama-3.2-1B-Instruct-FP8 \ --tensor-parallel-size 2Where:
--security-opt=label=disable- Disables SELinux label relabeling for volume mounts. Required for systems where SELinux is enabled. Without this option, the container might fail to start.
--shm-size=4g -p 8000:8000-
Specifies the shared memory size and port mapping. Increase
--shm-sizeto8GBif you experience shared memory issues. --userns=keep-id:uid=1001-
Maps the host UID to the effective UID of the vLLM process in the container. Alternatively, you can pass
--user=0, but this is less secure because it runs vLLM as root inside the container. --env "HUGGING_FACE_HUB_TOKEN=$HF_TOKEN"-
Specifies the Hugging Face API access token. Set and export
HF_TOKENwith your Hugging Face token. -v ./rhaii-cache:/opt/app-root/src/.cache:Z-
Mounts the cache directory with SELinux context. The
:Zsuffix is required for systems where SELinux is enabled. On Debian, Ubuntu, or Docker without SELinux, omit the:Zsuffix. --tensor-parallel-size 2- Specifies the number of GPUs to use for tensor parallelism. Set this value to match the number of available GPUs.
In a separate tab in your terminal, make a request to your model with the API.
curl -X POST -H "Content-Type: application/json" -d '{ "prompt": "What is the capital of France?", "max_tokens": 50 }' http://<your_server_ip>:8000/v1/completions | jqExample output
{ "id": "cmpl-b84aeda1d5a4485c9cb9ed4a13072fca", "object": "text_completion", "created": 1746555421, "model": "RedHatAI/Llama-3.2-1B-Instruct-FP8", "choices": [ { "index": 0, "text": " Paris.\nThe capital of France is Paris.", "logprobs": null, "finish_reason": "stop", "stop_reason": null, "prompt_logprobs": null } ], "usage": { "prompt_tokens": 8, "total_tokens": 18, "completion_tokens": 10, "prompt_tokens_details": null } }
Chapter 5. Serving and inferencing with Podman using AMD ROCm AI accelerators Copy linkLink copied to clipboard!
Serve and inference a large language model with Podman and Red Hat AI Inference Server running on AMD ROCm AI accelerators.
Prerequisites
- You have installed Podman or Docker.
- You are logged in as a user with sudo access.
-
You have access to
registry.redhat.ioand have logged in. - You have a Hugging Face account and have generated a Hugging Face access token.
You have access to a Linux server with data center grade AMD ROCm AI accelerators installed.
For AMD GPUs:
For more information about supported vLLM quantization schemes for accelerators, see Supported hardware.
Procedure
Open a terminal on your server host, and log in to
registry.redhat.io:$ podman login registry.redhat.ioPull the AMD ROCm image by running the following command:
$ podman pull registry.redhat.io/rhaii-early-access/vllm-rocm-rhel9:3.4.0-ea.1If your system has SELinux enabled, configure SELinux to allow device access:
$ sudo setsebool -P container_use_devices 1Create a volume and mount it into the container. Adjust the container permissions so that the container can use it.
$ mkdir -p rhaii-cache$ chmod g+rwX rhaii-cacheCreate or append your
HF_TOKENHugging Face token to theprivate.envfile. Source theprivate.envfile.$ echo "export HF_TOKEN=<your_HF_token>" > private.env$ source private.envStart the AI Inference Server container image.
For AMD ROCm accelerators:
Use
amd-smi static -ato verify that the container can access the host system GPUs:$ podman run -ti --rm --pull=newer \ --security-opt=label=disable \ --device=/dev/kfd --device=/dev/dri \ --group-add keep-groups \ --entrypoint="" \ registry.redhat.io/rhaii-early-access/vllm-rocm-rhel9:3.4.0-ea.1 \ amd-smi static -aWhere:
--group-add keep-groups-
Preserves the supplementary groups from the host user. On AMD systems, you must belong to both the
videoandrendergroups to access GPUs.
Start the container:
podman run --rm -it \ --device /dev/kfd --device /dev/dri \ --security-opt=label=disable \ --group-add keep-groups \ --shm-size=4GB -p 8000:8000 \ --env "HUGGING_FACE_HUB_TOKEN=$HF_TOKEN" \ --env "HF_HUB_OFFLINE=0" \ -v ./rhaii-cache:/opt/app-root/src/.cache \ registry.redhat.io/rhaii-early-access/vllm-rocm-rhel9:{rhaii-version} \ --model RedHatAI/Llama-3.2-1B-Instruct-FP8 \ --tensor-parallel-size 2Where:
--security-opt=label=disable- Disables SELinux label relabeling for volume mounts. Without this option, the container might fail to start.
--shm-size=4GB -p 8000:8000-
Specifies the shared memory size and port mapping. Increase
--shm-sizeto8GBif you experience shared memory issues. --tensor-parallel-size 2- Specifies the number of GPUs to use for tensor parallelism. Set this value to match the number of available GPUs.
Verification
In a separate tab in your terminal, make a request to the model with the API.
curl -X POST -H "Content-Type: application/json" -d '{ "prompt": "What is the capital of France?", "max_tokens": 50 }' http://<your_server_ip>:8000/v1/completions | jqExample output
{ "id": "cmpl-b84aeda1d5a4485c9cb9ed4a13072fca", "object": "text_completion", "created": 1746555421, "model": "RedHatAI/Llama-3.2-1B-Instruct-FP8", "choices": [ { "index": 0, "text": " Paris.\nThe capital of France is Paris.", "logprobs": null, "finish_reason": "stop", "stop_reason": null, "prompt_logprobs": null } ], "usage": { "prompt_tokens": 8, "total_tokens": 18, "completion_tokens": 10, "prompt_tokens_details": null } }
Chapter 6. Serving and inferencing language models with Podman using Google TPU AI accelerators Copy linkLink copied to clipboard!
Serve and inference a large language model with Podman or Docker and Red Hat AI Inference Server in a Google cloud VM that has Google TPU AI accelerators available.
Prerequisites
You have access to a Google cloud TPU VM with Google TPU AI accelerators configured. For more information, see:
- You have installed Podman or Docker.
- You are logged in as a user with sudo access.
-
You have access to the
registry.redhat.ioimage registry and have logged in. - You have a Hugging Face account and have generated a Hugging Face access token.
For more information about supported vLLM quantization schemes for accelerators, see Supported hardware.
Procedure
Open a terminal on your TPU server host, and log in to
registry.redhat.io:$ podman login registry.redhat.ioPull the Red Hat AI Inference Server image by running the following command:
$ podman pull registry.redhat.io/rhaii-early-access/vllm-tpu-rhel9:3.4.0-ea.1Optional: Verify that the TPUs are available in the host.
Open a shell prompt in the Red Hat AI Inference Server container. Run the following command:
$ podman run -it --net=host --privileged --rm --entrypoint /bin/bash registry.redhat.io/rhaii-early-access/vllm-tpu-rhel9:3.4.0-ea.1Verify system TPU access and basic operations by running the following Python code in the container shell prompt:
$ python3 -c " import jax import importlib.metadata try: devices = jax.devices() print(f'JAX devices available: {devices}') print(f'Number of TPU devices: {len(devices)}') tpu_version = importlib.metadata.version('tpu_inference') print(f'tpu-inference version: {tpu_version}') print('TPU is operational.') except Exception as e: print(f'TPU test failed: {e}') print('Check container image version and TPU device availability.') "Example output:
JAX devices available: [TpuDevice(id=0, process_index=0, coords=(0,0,0), core_on_chip=0), TpuDevice(id=1, process_index=0, coords=(1,0,0), core_on_chip=0), TpuDevice(id=2, process_index=0, coords=(0,1,0), core_on_chip=0), TpuDevice(id=3, process_index=0, coords=(1,1,0), core_on_chip=0)] Number of TPU devices: 4 tpu-inference version: 0.13.2 TPU is operational.Exit the shell prompt.
$ exit
Create a volume and mount it into the container. Adjust the container permissions so that the container can use it.
$ mkdir ./.cache/rhaii$ chmod g+rwX ./.cache/rhaiiAdd the
HF_TOKENHugging Face token to theprivate.envfile.$ echo "export HF_TOKEN=<huggingface_token>" > private.envAppend the
HF_HOMEvariable to theprivate.envfile.$ echo "export HF_HOME=./.cache/rhaii" >> private.envSource the
private.envfile.$ source private.envStart the AI Inference Server container image:
$ podman run --rm -it \ --name vllm-tpu \ --network=host \ --privileged \ -v /dev/shm:/dev/shm \ -e HF_TOKEN=$HF_TOKEN \ -e HF_HUB_OFFLINE=0 \ -v ./.cache/rhaii:/opt/app-root/src/.cache \ registry.redhat.io/rhaii-early-access/vllm-tpu-rhel9:3.4.0-ea.1 \ --model Qwen/Qwen2.5-1.5B-Instruct \ --tensor-parallel-size 1 \ --max-model-len=256 \ --host=0.0.0.0 \ --port=8000Where:
--tensor-parallel-size 1- Specifies the number of TPUs to use for tensor parallelism. Set this value to match the number of available TPUs.
--max-model-len=256- Specifies the maximum model context length. For optimal performance, set this value as low as your workload allows.
Verification
Check that the AI Inference Server server is up. Open a separate tab in your terminal, and make a model request with the API:
curl http://localhost:8000/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"model": "Qwen/Qwen2.5-1.5B-Instruct",
"messages": [
{"role": "user", "content": "Briefly, what colour is the wind?"}
],
"max_tokens": 50
}' | jq
+ The model returns a valid JSON response answering your question.
Chapter 7. Inference serving with Podman on IBM Power with IBM Spyre AI accelerators Copy linkLink copied to clipboard!
Serve and inference a large language model with Podman and Red Hat AI Inference Server running on IBM Power with IBM Spyre AI accelerators.
Prerequisites
- You have access to an IBM Power 11 server running RHEL 9.6 with IBM Spyre for Power AI accelerators installed.
- You are logged in as a user with sudo access.
- You have installed Podman.
-
You have access to
registry.redhat.ioand have logged in. - You have installed the Service Report tool. See IBM Power Systems service and productivity tools.
-
You have created a
sentientsecurity group and added your Spyre user to the group.
Procedure
Open a terminal on your server host, and log in to
registry.redhat.io:$ podman login registry.redhat.ioRun the
servicereportcommand to verify your IBM Spyre hardware:$ servicereport -r -p spyreExample output
servicereport 2.2.5 Spyre configuration checks PASS VFIO Driver configuration PASS User memlock configuration PASS sos config PASS sos package PASS VFIO udev rules configuration PASS User group configuration PASS VFIO device permission PASS VFIO kernel module loaded PASS VFIO module dep configuration PASS Memlock limit is set for the sentient group. Spyre user must be in the sentient group. To add run below command: sudo usermod -aG sentient <user> Example: sudo usermod -aG sentient abc Re-login as <user>.Pull the Red Hat AI Inference Server image by running the following command:
$ podman pull registry.redhat.io/rhaii-early-access/vllm-spyre:3.4.0-ea.1If your system has SELinux enabled, configure SELinux to allow device access:
$ sudo setsebool -P container_use_devices 1Use
lspci -vto verify that the container can access the host system IBM Spyre AI accelerators:$ podman run -it --rm --pull=newer \ --security-opt=label=disable \ --device=/dev/vfio \ --group-add keep-groups \ --entrypoint="lspci" \ registry.redhat.io/rhaii-early-access/vllm-spyre:3.4.0-ea.1Example output
0381:50:00.0 Processing accelerators: IBM Spyre Accelerator (rev 02) 0382:60:00.0 Processing accelerators: IBM Spyre Accelerator (rev 02) 0383:70:00.0 Processing accelerators: IBM Spyre Accelerator (rev 02) 0384:80:00.0 Processing accelerators: IBM Spyre Accelerator (rev 02)Create a volume to mount into the container and adjust the container permissions so that the container can use it.
$ mkdir -p ~/models && chmod g+rwX ~/modelsDownload the
granite-3.3-8b-instructmodel into themodels/folder. See Downloading models for more information.NoteAs an alternative to downloading models from Hugging Face, you can use validated Red Hat AI modelcar container images with a
3.0or later tag. For more information about using modelcar images, see Inference serving language models in OCI-compliant model containers.Gather the Spyre IDs for the
VLLM_AIU_PCIE_IDSvariable:$ lspciExample output
0381:50:00.0 Processing accelerators: IBM Spyre Accelerator (rev 02) 0382:60:00.0 Processing accelerators: IBM Spyre Accelerator (rev 02) 0383:70:00.0 Processing accelerators: IBM Spyre Accelerator (rev 02) 0384:80:00.0 Processing accelerators: IBM Spyre Accelerator (rev 02)Set the
SPYRE_IDSvariable:$ SPYRE_IDS="0381:50:00.0 0382:60:00.0 0383:70:00.0 0384:80:00.0"Start the AI Inference Server container. For example, deploy the granite-3.3-8b-instruct model configured for entity extraction inference serving:
podman run \ --device=/dev/vfio \ -v $HOME/models:/models \ -e AIU_PCIE_IDS="${SPYRE_IDS}" \ -e VLLM_SPYRE_USE_CB=1 \ --pids-limit 0 \ --userns=keep-id \ --group-add=keep-groups \ --memory 200G \ --shm-size 64G \ -p 8000:8000 \ registry.redhat.io/rhaii-early-access/vllm-spyre:3.4.0-ea.1 \ --model /models/granite-3.3-8b-instruct \ -tp 4 \ --max-model-len 32768 \ --max-num-seqs 32
Verification
In a separate tab in your terminal, make a request to the model with the API.
curl -X POST -H "Content-Type: application/json" -d '{ "model": "/models/granite-3.3-8b-instruct" "prompt": "What is the capital of France?", "max_tokens": 50 }' http://<your_server_ip>:8000/v1/completions | jqExample output
{ "id": "cmpl-b94beda1d5a4485c9cb9ed4a13072fca", "object": "text_completion", "created": 1746555421, "choices": [ { "index": 0, "text": " Paris.\nThe capital of France is Paris.", "logprobs": null, "finish_reason": "stop", "stop_reason": null, "prompt_logprobs": null } ], "usage": { "prompt_tokens": 8, "total_tokens": 18, "completion_tokens": 10, "prompt_tokens_details": null } }
7.1. Recommended model inference settings for IBM Power with IBM Spyre AI accelerators Copy linkLink copied to clipboard!
The following are the recommended model and AI Inference Server inference serving settings for IBM Power systems with IBM Spyre AI accelerators.
| Model | Batch size | Max input context size | Max output context size | Number of cards per container |
|---|---|---|---|---|
| granite3.3-8b-instruct | 16 | 3K | 3K | 1 |
| Model | Batch size | Max input context size | Max output context size | Number of cards per container |
|---|---|---|---|---|
| Up to 256 | 512 | Vector of size 768 | 1 |
| Up to 256 | 512 | Vector of size 384 | 1 |
| Model | Batch size | Max input context size | Max output context size | Number of cards per container |
|---|---|---|---|---|
| granite3.3-8b-instruct | 32 | 4K | 4K | 4 |
| 16 | 8K | 8K | 4 | |
| 8 | 16K | 16K | 4 | |
| 4 | 32K | 32K | 4 |
7.2. Example inference serving configurations for IBM Spyre AI accelerators on IBM Power Copy linkLink copied to clipboard!
The following examples describe common Red Hat AI Inference Server workloads on IBM Spyre AI accelerators and IBM Power.
- Entity extraction
Select a single Spyre card ID with the output from the
lspcicommand, for example:$ SPYRE_IDS="0381:50:00.0"Podman entity extraction example
$ podman run -d \ --device=/dev/vfio \ --name vllm-api \ -v $HOME/models:/models:Z \ -e VLLM_AIU_PCIE_IDS="$SPYRE_IDS" \ -e VLLM_SPYRE_USE_CB=1 \ --pids-limit 0 \ --userns=keep-id \ --group-add=keep-groups \ --memory 100GB \ --shm-size 64GB \ -p 8000:8000 \ registry.redhat.io/rhaii-early-access/vllm-spyre:3.4.0-ea.1 \ --enable-prefix-caching \ --model /models/granite-3.3-8b-instruct \ -tp 1 \ --max-model-len 3072 \ --max-num-seqs 16- RAG inference serving
Select 4 Spyre card IDs with the output from the
lspcicommand, for example:$ SPYRE_IDS="0381:50:00.0 0382:60:00.0 0383:70:00.0 0384:80:00.0"Podman RAG inference serving example
$ podman run -d \ --device=/dev/vfio \ --name vllm-api \ -v $HOME/models:/models:Z \ -e VLLM_AIU_PCIE_IDS="$SPYRE_IDS" \ -e VLLM_MODEL_PATH=/models/granite-3.3-8b-instruct \ -e VLLM_SPYRE_USE_CB=1 \ --pids-limit 0 \ --userns=keep-id \ --group-add=keep-groups \ --memory 200GB \ --shm-size 64GB \ -p 8000:8000 \ registry.redhat.io/rhaii-early-access/vllm-spyre:3.4.0-ea.1 \ --enable-prefix-caching \ --model /models/granite-3.3-8b-instruct \ -tp 4 \ --max-model-len 32768 \ --max-num-seqs 32- RAG embedding
Select a single Spyre card ID with the output from the
lspcicommand, for example:$ SPYRE_IDS="0384:80:00.0"Podman RAG embedding inference serving example
$ podman run -d \ --device=/dev/vfio \ --name vllm-api \ -v $HOME/models:/models:Z \ -e VLLM_AIU_PCIE_IDS="$SPYRE_IDS" \ -e VLLM_MODEL_PATH=/models/granite-embedding-125m-english \ -e VLLM_SPYRE_USE_CHUNKED_PREFILL=0 \ -e VLLM_SPYRE_WARMUP_PROMPT_LENS=64 \ -e VLLM_SPYRE_WARMUP_BATCH_SIZES=64 \ --pids-limit 0 \ --userns=keep-id \ --group-add=keep-groups \ --memory 200GB \ --shm-size 64GB \ -p 8000:8000 \ registry.redhat.io/rhaii-early-access/vllm-spyre:3.4.0-ea.1 \ --model /models/granite-embedding-125m-english \ -tp 1- Re-ranker inference serving
Select a single Spyre AI accelerator card ID with the output from the
lspcicommand, for example:$ SPYRE_IDS="0384:80:00.0"Podman re-ranker inference serving example
$ podman run -d \ --device=/dev/vfio \ --name vllm-api \ -v $HOME/models:/models:Z \ -e VLLM_AIU_PCIE_IDS="$SPYRE_IDS" \ -e VLLM_MODEL_PATH=/models/bge-reranker-v2-m3 \ -e VLLM_SPYRE_USE_CHUNKED_PREFILL=0 \ -e VLLM_SPYRE_WARMUP_PROMPT_LENS=1024 \ -e VLLM_SPYRE_WARMUP_BATCH_SIZES=4 \ --pids-limit 0 \ --userns=keep-id \ --group-add=keep-groups \ --memory 200GB \ --shm-size 64GB \ -p 8000:8000 \ registry.redhat.io/rhaii-early-access/vllm-spyre:3.4.0-ea.1 \ --model /models/bge-reranker-v2-m3 \ -tp 1
Chapter 8. Inference serving with Podman on IBM Z with IBM Spyre AI accelerators Copy linkLink copied to clipboard!
Serve and inference a large language model with Podman and Red Hat AI Inference Server running on IBM Z with IBM Spyre AI accelerators.
Prerequisites
- You have access to an IBM Z (s390x) server running RHEL 9.6 with IBM Spyre for Z AI accelerators installed.
- You are logged in as a user with sudo access.
- You have installed Podman.
-
You have access to
registry.redhat.ioand have logged in. - You have a Hugging Face account and have generated a Hugging Face access token.
IBM Spyre AI accelerator cards support FP16 format model weights only. For compatible models, the Red Hat AI Inference Server inference engine automatically converts weights to FP16 at startup. No additional configuration is needed.
Procedure
Open a terminal on your server host, and log in to
registry.redhat.io:$ podman login registry.redhat.ioPull the Red Hat AI Inference Server image by running the following command:
$ podman pull registry.redhat.io/rhaii-early-access/vllm-spyre:3.4.0-ea.1If your system has SELinux enabled, configure SELinux to allow device access:
$ sudo setsebool -P container_use_devices 1Use
lspci -vto verify that the container can access the host system IBM Spyre AI accelerators:$ podman run -it --rm --pull=newer \ --security-opt=label=disable \ --device=/dev/vfio \ --group-add keep-groups \ --entrypoint="lspci" \ registry.redhat.io/rhaii-early-access/vllm-spyre:3.4.0-ea.1Example output
0381:50:00.0 Processing accelerators: IBM Spyre Accelerator (rev 02) 0382:60:00.0 Processing accelerators: IBM Spyre Accelerator (rev 02) 0383:70:00.0 Processing accelerators: IBM Spyre Accelerator (rev 02) 0384:80:00.0 Processing accelerators: IBM Spyre Accelerator (rev 02)Create a volume to mount into the container and adjust the container permissions so that the container can use it.
$ mkdir -p ~/models && chmod g+rwX ~/modelsDownload the
granite-3.3-8b-instructmodel into themodels/folder. See Downloading models for more information.NoteAs an alternative to downloading models from Hugging Face, you can use validated Red Hat AI modelcar container images with a
3.0or later tag. For more information about using modelcar images, see Inference serving language models in OCI-compliant model containers.Gather the IOMMU group IDs for the available Spyre devices:
$ lspciExample output
0000:00:00.0 Processing accelerators: IBM Spyre Accelerator Virtual Function (rev 02) 0001:00:00.0 Processing accelerators: IBM Spyre Accelerator Virtual Function (rev 02) 0002:00:00.0 Processing accelerators: IBM Spyre Accelerator Virtual Function (rev ff) 0003:00:00.0 Processing accelerators: IBM Spyre Accelerator Virtual Function (rev 02)Each line begins with the PCI device address, for example,
0000:00:00.0.Use the PCI address to determine the IOMMU group ID for the required Spyre card, for example:
$ readlink /sys/bus/pci/devices/<PCI_ADDRESS>/iommu_groupExample output
../../../kernel/iommu_groups/0The IOMMU group ID (0) is the trailing number in the
readlinkoutput.Repeat for each required Spyre card.
Set
IOMMU_GROUP_IDvariables for the required Spyre cards using thereadlinkoutput. For example:IOMMU_GROUP_ID0=0 IOMMU_GROUP_ID1=1 IOMMU_GROUP_ID2=2 IOMMU_GROUP_ID3=3Start the AI Inference Server container, passing in the IOMMU group ID variables for the required Spyre devices. For example, deploy the granite-3.3-8b-instruct model configured for entity extraction across 4 Spyre devices:
podman run \ --device /dev/vfio/vfio \ --device /dev/vfio/${IOMMU_GROUP_ID0}:/dev/vfio/${IOMMU_GROUP_ID0} \ --device /dev/vfio/${IOMMU_GROUP_ID1}:/dev/vfio/${IOMMU_GROUP_ID1} \ --device /dev/vfio/${IOMMU_GROUP_ID2}:/dev/vfio/${IOMMU_GROUP_ID2} \ --device /dev/vfio/${IOMMU_GROUP_ID3}:/dev/vfio/${IOMMU_GROUP_ID3} \ -v $HOME/models:/models:Z \ --pids-limit 0 \ --userns=keep-id \ --group-add=keep-groups \ --memory 200G \ --shm-size 64G \ -p 8000:8000 \ -e VLLM_DT_CHUNK_LEN=512 \ registry.redhat.io/rhaii-early-access/vllm-spyre:3.4.0-ea.1 \ --model /models/granite-3.3-8b-instruct \ -tp 4 \ --max-model-len 32768 \ --max-num-seqs 32 \ --enable-prefix-caching
Verification
In a separate tab in your terminal, make a request to the model with the API.
curl -X POST -H "Content-Type: application/json" -d '{ "model": "/models/granite-3.3-8b-instruct", "prompt": "What is the capital of France?", "max_tokens": 50 }' http://<your_server_ip>:8000/v1/completions | jqExample output
{ "id": "cmpl-7c81cd00ccd04237ac8b5119e86b32a5", "object": "text_completion", "created": 1764665204, "model": "/models/granite-3.3-8b-instruct", "choices": [ { "index": 0, "text": "\nThe answer is Paris. Paris is the capital and most populous city of France, located in the northern part of the country. It is renowned for its history, culture, fashion, and art, attracting", "logprobs": null, "finish_reason": "length", "stop_reason": null, "token_ids": null, "prompt_logprobs": null, "prompt_token_ids": null } ], "service_tier": null, "system_fingerprint": null, "usage": { "prompt_tokens": 7, "total_tokens": 57, "completion_tokens": 50, "prompt_tokens_details": null }, "kv_transfer_params": null }
Chapter 9. Serving and inferencing language models with Podman using AWS Trainium and Inferentia AI accelerators Copy linkLink copied to clipboard!
Serve and inference a large language model with Podman or Docker and Red Hat AI Inference Server on an AWS cloud instance that has AWS Trainium or Inferentia AI accelerators configured.
AWS Inferentia and AWS Trainium are custom-designed machine learning chips from Amazon Web Services (AWS). Red Hat AI Inference Server integrates with these accelerators through the AWS Neuron SDK, providing a path to deploy vLLM-based inference workloads on AWS cloud infrastructure.
AWS Trainium and Inferentia support is a Technology Preview feature only. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process.
For more information about the support scope of Red Hat Technology Preview features, see Technology Preview Features Support Scope.
Prerequisites
- You have access to an AWS Inf2, Trn1, Trn1n, or Trn2 instance with AWS Neuron drivers configured. See Neuron setup guide.
- You have installed Podman or Docker.
- You are logged in as a user that has sudo access.
-
You have access to the
registry.redhat.ioimage registry. - You have a Hugging Face account and have generated a Hugging Face access token.
Procedure
Open a terminal on your AWS host, and log in to
registry.redhat.io:$ podman login registry.redhat.ioPull the Red Hat AI Inference Server image for Neuron by running the following command:
$ podman pull registry.redhat.io/rhaii-early-access/vllm-neuron-rhel9:3.4.0-ea.1Optional: Verify that the Neuron drivers and devices are available on the host.
Run
neuron-lsto verify that Neuron drivers are installed and to view detailed information about the Neuron hardware:$ neuron-lsExample output
instance-type: trn1.2xlarge instance-id: i-0b29616c0f73dc323 +--------+--------+----------+--------+--------------+----------+------+ | NEURON | NEURON | NEURON | NEURON | PCI | CPU | NUMA | | DEVICE | CORES | CORE IDS | MEMORY | BDF | AFFINITY | NODE | +--------+--------+----------+--------+--------------+----------+------+ | 0 | 2 | 0-1 | 32 GB | 0000:00:1e.0 | 0-7 | -1 | +--------+--------+----------+--------+--------------+----------+------+Note the number of Neuron cores available. Use this information to set
--tensor-parallel-sizeargument when starting the container.List the Neuron devices:
$ ls /dev/neuron*Example output
/dev/neuron0
Create a volume for mounting into the container and adjust the permissions so that the container can use it:
$ mkdir -p ./.cache/rhaii && chmod g+rwX ./.cache/rhaiiAdd the
HF_TOKENHugging Face token to theprivate.envfile.$ echo "export HF_TOKEN=<huggingface_token>" > private.envAppend the
HF_HOMEvariable to theprivate.envfile.$ echo "export HF_HOME=./.cache/rhaii" >> private.envSource the
private.envfile.$ source private.envStart the AI Inference Server container image:
$ sudo podman run -it --rm \ -e HUGGING_FACE_HUB_TOKEN=$HF_TOKEN \ -e HF_HUB_OFFLINE=0 \ --network=host \ --device=/dev/neuron0 \ -p 8000:8000 \ -v $HOME/.cache/rhaii:/root/.cache/huggingface \ -v ./.cache/rhaii:/opt/app-root/src/.cache:Z \ registry.redhat.io/rhaii-early-access/vllm-neuron-rhel9:3.4.0-ea.1 \ --model Qwen/Qwen2.5-1.5B-Instruct \ --max-model-len 4096 \ --max-num-seqs 1 \ --no-enable-prefix-caching \ --port 8000 \ --tensor-parallel-size 2 \ --additional-config '{ "override_neuron_config": { "async_mode": false } }'--device=/dev/neuron0- Map the required Neuron device. Adjust based on your model requirements and available Neuron memory.
--no-enable-prefix-caching- Disable prefix caching for Neuron hardware.
--tensor-parallel-size 2-
Set
--tensor-parallel-sizeto match the number of neuron cores being used. --additional-config '{ "override_neuron_config": { "async_mode": false } }'-
The
--additional-configparameter passes Neuron-specific configuration. Settingasync_modetofalseis recommended for stability.
Verification
Check that the AI Inference Server server is up. Open a separate tab in your terminal, and make a model request with the API:
$ curl http://localhost:8000/v1/chat/completions \ -H "Content-Type: application/json" \ -d '{ "model": "Qwen/Qwen2.5-1.5B-Instruct", "messages": [ {"role": "user", "content": "Briefly, what color is the wind?"} ], "max_tokens": 50 }' | jqExample output
{ "id": "chatcmpl-abc123def456", "object": "chat.completion", "created": 1755268559, "model": "Qwen/Qwen2.5-1.5B-Instruct", "choices": [ { "index": 0, "message": { "role": "assistant", "content": "The wind is typically associated with the color white or grey, as it can carry dust, sand, or other particles. However, it is not a color in the traditional sense.", "refusal": null, "annotations": null, "audio": null, "function_call": null, "tool_calls": [], "reasoning_content": null }, "logprobs": null, "finish_reason": "stop", "stop_reason": null } ], "service_tier": null, "system_fingerprint": null, "usage": { "prompt_tokens": 38, "total_tokens": 75, "completion_tokens": 37, "prompt_tokens_details": null }, "prompt_logprobs": null, "kv_transfer_params": null }
Chapter 10. Serving and inferencing with Podman using CPU (x86_64 AVX2) Copy linkLink copied to clipboard!
Serve and inference a large language model with Podman and Red Hat AI Inference Server running on x86_64 CPUs with AVX2 instruction set support.
With CPU-only inference, you can run Red Hat AI Inference Server workloads on x86_64 CPUs without requiring GPU hardware. This feature provides a cost-effective option for development, testing, and small-scale deployments using smaller language models.
Inference serving with AI Inference Server on x86_64 AVX2 CPU is a Technology Preview feature only. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process.
For more information about the support scope of Red Hat Technology Preview features, see Technology Preview Features Support Scope.
AVX512 instruction set support is planned for a future release.
Prerequisites
- You have installed Podman or Docker.
- You are logged in as a user with sudo access.
-
You have access to
registry.redhat.ioand have logged in. - You have a Hugging Face account and have generated a Hugging Face access token.
You have access to a Linux server with an x86_64 CPU that supports the AVX2 instruction set:
- Intel Haswell (2013) or newer processors
- AMD Excavator (2015) or newer processors
- You have a minimum of 16GB system RAM. 32GB or more is recommended for larger models.
CPU inference is optimized for smaller models, typically under 3 billion parameters. For larger models or production workloads requiring higher throughput, consider using GPU acceleration.
Procedure
Open a terminal on your server host, and log in to
registry.redhat.io:$ podman login registry.redhat.ioPull the CPU inference image by running the following command:
$ podman pull registry.redhat.io/rhaii-early-access/vllm-cpu-rhel9:3.4.0-ea.1Create a volume and mount it into the container. Adjust the container permissions so that the container can use it.
$ mkdir -p rhaii-cache && chmod g+rwX rhaii-cacheCreate or append your
HF_TOKENHugging Face token to theprivate.envfile. Source theprivate.envfile.$ echo "export HF_TOKEN=<your_HF_token>" > private.env$ source private.envVerify that your CPU supports the AVX2 instruction set:
$ grep -q avx2 /proc/cpuinfo && echo "AVX2 supported" || echo "AVX2 not supported"ImportantIf your CPU does not support AVX2, you cannot use CPU inference with Red Hat AI Inference Server.
Start the AI Inference Server container image.
$ podman run --rm -it \ --security-opt=label=disable \ --shm-size=4g -p 8000:8000 \ --userns=keep-id:uid=1001 \ --env "HUGGING_FACE_HUB_TOKEN=$HF_TOKEN" \ --env "HF_HUB_OFFLINE=0" \ --env "VLLM_CPU_KVCACHE_SPACE=4" \ -v ./rhaii-cache:/opt/app-root/src/.cache:Z \ registry.redhat.io/rhaii-early-access/vllm-cpu-rhel9:3.4.0-ea.1 \ --model TinyLlama/TinyLlama-1.1B-Chat-v1.0-
--security-opt=label=disable: Disables SELinux label relabeling for volume mounts. Required for systems where SELinux is enabled. Without this option, the container might fail to start. -
--shm-size=4g -p 8000:8000: Specifies the shared memory size and port mapping. Increase--shm-sizeto8GBif you experience shared memory issues. -
--userns=keep-id:uid=1001: Maps the host UID to the effective UID of the vLLM process in the container. Alternatively, you can pass--user=0, but this is less secure because it runs vLLM as root inside the container. -
--env "HUGGING_FACE_HUB_TOKEN=$HF_TOKEN": Specifies the Hugging Face API access token. Set and exportHF_TOKENwith your Hugging Face token. -
--env "VLLM_CPU_KVCACHE_SPACE=4": Allocates 4GB for the CPU key-value cache. Increase this value for larger models or longer context lengths. The default is 4GB. -
-v ./rhaii-cache:/opt/app-root/src/.cache:Z: Mounts the cache directory with SELinux context. The:Zsuffix is required for systems where SELinux is enabled. On Debian, Ubuntu, or Docker without SELinux, omit the:Zsuffix. -
--model TinyLlama/TinyLlama-1.1B-Chat-v1.0: Specifies the Hugging Face model to serve. For CPU inference, use smaller models (under 3B parameters) for optimal performance.
-
Verification
In a separate tab in your terminal, make a request to the model with the API.
curl -X POST -H "Content-Type: application/json" -d '{ "prompt": "What is the capital of France?", "max_tokens": 50 }' http://<your_server_ip>:8000/v1/completions | jqThe model returns a valid JSON response answering your question.
Chapter 11. Validating Red Hat AI Inference Server benefits using key metrics Copy linkLink copied to clipboard!
Use the following metrics to evaluate the performance of the LLM model being served with AI Inference Server:
- Time to first token (TTFT): The time from when a request is sent to when the first token of the response is received.
- Time per output token (TPOT): The average time it takes to generate each token after the first one.
- Latency: The total time required to generate the full response.
- Throughput: The total number of output tokens the model can produce at the same time across all users and requests.
Complete the procedure below to run a benchmark test that shows how AI Inference Server, and other inference servers, perform according to these metrics.
Prerequisites
- AI Inference Server container image
- GitHub account
- Python 3.9 or higher
Procedure
On your host system, start an AI Inference Server container and serve a model.
$ podman run --rm -it --device nvidia.com/gpu=all \ --shm-size=4GB -p 8000:8000 \ --env "HUGGING_FACE_HUB_TOKEN=$HF_TOKEN" \ --env "HF_HUB_OFFLINE=0" \ -v ./rhaii-cache:/opt/app-root/src/.cache \ --security-opt=label=disable \ registry.redhat.io/rhaii-early-access/vllm-cuda-rhel9:3.4.0-ea.1 \ --model RedHatAI/Llama-3.2-1B-Instruct-FP8In a separate terminal tab, install the benchmark tool dependencies.
$ pip install vllm pandas datasetsClone the vLLM Git repository:
$ git clone https://github.com/vllm-project/vllm.gitRun the
./vllm/benchmarks/benchmark_serving.pyscript.$ python vllm/benchmarks/benchmark_serving.py --backend vllm --model RedHatAI/Llama-3.2-1B-Instruct-FP8 --num-prompts 100 --dataset-name random --random-input 1024 --random-output 512 --port 8000
Verification
The results show how AI Inference Server performs according to key server metrics:
============ Serving Benchmark Result ============
Successful requests: 100
Benchmark duration (s): 4.61
Total input tokens: 102300
Total generated tokens: 40493
Request throughput (req/s): 21.67
Output token throughput (tok/s): 8775.85
Total Token throughput (tok/s): 30946.83
---------------Time to First Token----------------
Mean TTFT (ms): 193.61
Median TTFT (ms): 193.82
P99 TTFT (ms): 303.90
-----Time per Output Token (excl. 1st token)------
Mean TPOT (ms): 9.06
Median TPOT (ms): 8.57
P99 TPOT (ms): 13.57
---------------Inter-token Latency----------------
Mean ITL (ms): 8.54
Median ITL (ms): 8.49
P99 ITL (ms): 13.14
==================================================
Try changing the parameters of this benchmark and running it again. Notice how vllm as a backend compares to other options. Throughput should be consistently higher, while latency should be lower.
-
Other options for
--backendare:tgi,lmdeploy,deepspeed-mii,openai, andopenai-chat -
Other options for
--dataset-nameare:sharegpt,burstgpt,sonnet,random,hf
Additional resources
- vLLM documentation
- LLM Inference Performance Engineering: Best Practices, by Mosaic AI Research, which explains metrics such as throughput and latency
Chapter 12. Troubleshooting Copy linkLink copied to clipboard!
The following troubleshooting information for Red Hat AI Inference Server 3.4.0-ea.1 describes common problems related to model loading, memory, model response quality, networking, and GPU drivers. Where available, workarounds for common issues are described.
Most common issues in vLLM relate to installation, model loading, memory management, and GPU communication. Most problems can be resolved by using a correctly configured environment, ensuring compatible hardware and software versions, and following the recommended configuration practices.
For persistent issues, export VLLM_LOGGING_LEVEL=DEBUG to enable debug logging and then check the logs.
$ export VLLM_LOGGING_LEVEL=DEBUG
12.1. Model loading errors Copy linkLink copied to clipboard!
When you run the Red Hat AI Inference Server container image without specifying a user namespace, an unrecognized model error is returned.
podman run --rm -it \ --device nvidia.com/gpu=all \ --security-opt=label=disable \ --shm-size=4GB -p 8000:8000 \ --env "HUGGING_FACE_HUB_TOKEN=$HF_TOKEN" \ --env "HF_HUB_OFFLINE=0" \ -v ./rhaii-cache:/opt/app-root/src/.cache \ registry.redhat.io/rhaii-early-access/vllm-cuda-rhel9:3.4.0-ea.1 \ --model RedHatAI/Llama-3.2-1B-Instruct-FP8Example output
ValueError: Unrecognized model in RedHatAI/Llama-3.2-1B-Instruct-FP8. Should have a model_type key in its config.jsonTo resolve this error, pass
--userns=keep-id:uid=1001as the first Podman parameter to ensure that the container runs with the root user.podman run --rm -it \ --userns=keep-id:uid=1001 \ --device nvidia.com/gpu=all \ --security-opt=label=disable \ --shm-size=4GB -p 8000:8000 \ --env "HUGGING_FACE_HUB_TOKEN=$HF_TOKEN" \ --env "HF_HUB_OFFLINE=0" \ -v ./rhaii-cache:/opt/app-root/src/.cache \ registry.redhat.io/{rhaii-registry-namespace}/vllm-cuda-rhel9:{rhaiis-version} \ --model RedHatAI/Llama-3.2-1B-Instruct-FP8Sometimes when Red Hat AI Inference Server downloads the model, the download fails or gets stuck. To prevent the model download from hanging, first download the model using the
huggingface-cli. For example:$ huggingface-cli download <MODEL_ID> --local-dir <DOWNLOAD_PATH>When serving the model, pass the local model path to vLLM to prevent the model from being downloaded again.
When Red Hat AI Inference Server loads a model from disk, the process sometimes hangs. Large models consume memory, and if memory runs low, the system slows down as it swaps data between RAM and disk. Slow network file system speeds or a lack of available memory can trigger excessive swapping. This can happen in clusters where file systems are shared between cluster nodes.
Where possible, store the model in a local disk to prevent slow down during model loading. Ensure that the system has sufficient CPU memory available.
Ensure that your system has enough CPU capacity to handle the model.
Sometimes, Red Hat AI Inference Server fails to inspect the model. Errors are reported in the log. For example:
#... File "vllm/model_executor/models/registry.py", line xxx, in \_raise_for_unsupported raise ValueError( ValueError: Model architectures [''] failed to be inspected. Please check the logs for more details.The error occurs when vLLM fails to import the model file, which is usually related to missing dependencies or outdated binaries in the vLLM build.
Some model architectures are not supported. Refer to the list of Red Hat AI validated models. For example, the following errors indicate that the model you are trying to use is not supported:
Traceback (most recent call last): #... File "vllm/model_executor/models/registry.py", line xxx, in inspect_model_cls for arch in architectures: TypeError: 'NoneType' object is not iterable#... File "vllm/model_executor/models/registry.py", line xxx, in \_raise_for_unsupported raise ValueError( ValueError: Model architectures [''] are not supported for now. Supported architectures: #...NoteSome architectures such as
DeepSeekV2VLrequire the architecture to be explicitly specified using the--hf_overridesflag, for example:--hf_overrides '{\"architectures\": [\"DeepseekVLV2ForCausalLM\"]}Sometimes a runtime error occurs for certain hardware when you load 8-bit floating point (FP8) models. FP8 requires GPU hardware acceleration. Errors occur when you load FP8 models like
deepseek-r1or models tagged with theF8_E4M3tensor type. For example:triton.compiler.errors.CompilationError: at 1:0: def \_per_token_group_quant_fp8( \^ ValueError("type fp8e4nv not supported in this architecture. The supported fp8 dtypes are ('fp8e4b15', 'fp8e5')") [rank0]:[W502 11:12:56.323757996 ProcessGroupNCCL.cpp:1496] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())NoteReview Getting started to ensure your specific accelerator is supported. Accelerators that are currently supported for FP8 models include:
Sometimes when serving a model a runtime error occurs that is related to the host system. For example, you might see errors in the log like this:
INFO 05-07 19:15:17 [config.py:1901] Chunked prefill is enabled with max_num_batched_tokens=2048. OMP: Error #179: Function Can't open SHM failed: OMP: System error #0: Success Traceback (most recent call last): File "/opt/app-root/bin/vllm", line 8, in <module> sys.exit(main()) .......................... raise RuntimeError("Engine core initialization failed. " RuntimeError: Engine core initialization failed. See root cause above.You can work around this issue by passing the
--shm-size=2gargument when startingvllm.
12.2. Memory optimization Copy linkLink copied to clipboard!
- If the model is too large to run with a single GPU, you will get out-of-memory (OOM) errors. Use memory optimization options such as quantization, tensor parallelism, or reduced precision to reduce the memory consumption. For more information, see Conserving memory.
12.3. Generated model response quality Copy linkLink copied to clipboard!
In some scenarios, the quality of the generated model responses might deteriorate after an update.
Default sampling parameters source have been updated in newer versions. For vLLM version 0.8.4 and higher, the default sampling parameters come from the
generation_config.jsonfile that is provided by the model creator. In most cases, this should lead to higher quality responses, because the model creator is likely to know which sampling parameters are best for their model. However, in some cases the defaults provided by the model creator can lead to degraded performance.If you experience this problem, try serving the model with the old defaults by using the
--generation-config vllmserver argument.ImportantIf applying the
--generation-config vllmserver argument improves the model output, continue to use the vLLM defaults and petition the model creator on Hugging Face to update their defaultgeneration_config.jsonso that it produces better quality generations.
12.4. CUDA accelerator errors Copy linkLink copied to clipboard!
You might experience a
self.graph.replay()error when running a model using CUDA accelerators.If vLLM crashes and the error trace captures the error somewhere around the
self.graph.replay()method in thevllm/worker/model_runner.pymodule, this is most likely a CUDA error that occurs inside theCUDAGraphclass.To identify the particular CUDA operation that causes the error, add the
--enforce-eagerserver argument to thevllmcommand line to disableCUDAGraphoptimization and isolate the problematic CUDA operation.You might experience accelerator and CPU communication problems that are caused by incorrect hardware or driver settings.
NVIDIA Fabric Manager is required for multi-GPU systems for some types of NVIDIA GPUs. The
nvidia-fabricmanagerpackage and associated systemd service might not be installed or the package might not be running.Run the diagnostic Python script to check whether the NVIDIA Collective Communications Library (NCCL) and Gloo library components are communicating correctly.
On an NVIDIA system, check the fabric manager status by running the following command:
$ systemctl status nvidia-fabricmanagerOn successfully configured systems, the service should be active and running with no errors.
-
Running vLLM with tensor parallelism enabled and setting
--tensor-parallel-sizeto be greater than 1 on NVIDIA Multi-Instance GPU (MIG) hardware causes anAssertionErrorduring the initial model loading or shape checking phase. This typically occurs as one of the first errors when starting vLLM.
12.5. Networking errors Copy linkLink copied to clipboard!
You might experience network errors with complicated network configurations.
To troubleshoot network issues, search the logs for DEBUG statements where an incorrect IP address is listed, for example:
DEBUG 06-10 21:32:17 parallel_state.py:88] world_size=8 rank=0 local_rank=0 distributed_init_method=tcp://<incorrect_ip_address>:54641 backend=ncclTo correct the issue, set the correct IP address with the
VLLM_HOST_IPenvironment variable, for example:$ export VLLM_HOST_IP=<correct_ip_address>Specify the network interface that is tied to the IP address for NCCL and Gloo:
$ export NCCL_SOCKET_IFNAME=<your_network_interface>$ export GLOO_SOCKET_IFNAME=<your_network_interface>
12.6. Python multiprocessing errors Copy linkLink copied to clipboard!
You might experience Python multiprocessing warnings or runtime errors. This can be caused by code that is not properly structured for Python multiprocessing. The following is an example console warning:
WARNING 12-11 14:50:37 multiproc_worker_utils.py:281] CUDA was previously initialized. We must use the `spawn` multiprocessing start method. Setting VLLM_WORKER_MULTIPROC_METHOD to 'spawn'.The following is an example Python runtime error:
RuntimeError: An attempt has been made to start a new process before the current process has finished its bootstrapping phase. This probably means that you are not using fork to start your child processes and you have forgotten to use the proper idiom in the main module: if __name__ = "__main__": freeze_support() ... The "freeze_support()" line can be omitted if the program is not going to be frozen to produce an executable. To fix this issue, refer to the "Safe importing of main module" section in https://docs.python.org/3/library/multiprocessing.htmlTo resolve the runtime error, update your Python code to guard the usage of
vllmbehind anif__name__ = "__main__":block, for example:if __name__ = "__main__": import vllm llm = vllm.LLM(...)
12.7. GPU driver or device pass-through issues Copy linkLink copied to clipboard!
When you run the Red Hat AI Inference Server container image, sometimes it is unclear whether device pass-through errors are being caused by GPU drivers or tools such as the NVIDIA Container Toolkit.
Check that the NVIDIA Container toolkit that is installed on the host machine can see the host GPUs:
$ nvidia-ctk cdi listExample output
#... nvidia.com/gpu=GPU-0fe9bb20-207e-90bf-71a7-677e4627d9a1 nvidia.com/gpu=GPU-10eff114-f824-a804-e7b7-e07e3f8ebc26 nvidia.com/gpu=GPU-39af96b4-f115-9b6d-5be9-68af3abd0e52 nvidia.com/gpu=GPU-3a711e90-a1c5-3d32-a2cd-0abeaa3df073 nvidia.com/gpu=GPU-6f5f6d46-3fc1-8266-5baf-582a4de11937 nvidia.com/gpu=GPU-da30e69a-7ba3-dc81-8a8b-e9b3c30aa593 nvidia.com/gpu=GPU-dc3c1c36-841b-bb2e-4481-381f614e6667 nvidia.com/gpu=GPU-e85ffe36-1642-47c2-644e-76f8a0f02ba7 nvidia.com/gpu=allEnsure that the NVIDIA accelerator configuration has been created on the host machine:
$ sudo nvidia-ctk cdi generate --output=/etc/cdi/nvidia.yamlCheck that the Red Hat AI Inference Server container can access NVIDIA GPUs on the host by running the following command:
$ podman run --rm -it --security-opt=label=disable --device nvidia.com/gpu=all nvcr.io/nvidia/cuda:12.4.1-base-ubi9 nvidia-smiExample output
+-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 570.124.06 Driver Version: 570.124.06 CUDA Version: 12.8 | |-----------------------------------------+------------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 NVIDIA A100-SXM4-80GB Off | 00000000:08:01.0 Off | 0 | | N/A 32C P0 64W / 400W | 1MiB / 81920MiB | 0% Default | | | | Disabled | +-----------------------------------------+------------------------+----------------------+ | 1 NVIDIA A100-SXM4-80GB Off | 00000000:08:02.0 Off | 0 | | N/A 29C P0 63W / 400W | 1MiB / 81920MiB | 0% Default | | | | Disabled | +-----------------------------------------+------------------------+----------------------+ +-----------------------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=========================================================================================| | No running processes found | +-----------------------------------------------------------------------------------------+
12.8. Troubleshooting IBM Power issues Copy linkLink copied to clipboard!
If you are unable to access the model data from the AI Inference Server container, complete the following steps:
-
Verify that the
/modelsfolder mapping to the container is correct - Review the host SELinux settings
Ensure that you have applied appropriate permissions on the
$HOME/modelsfolder, for example:$ chmod -R 755 $HOME/modelsEnsure that you are using the
:Zoption for the Podman volume mounts:$ podman run -d --device=/dev/vfio \ -v $HOME/models:/models:Z \ # ...-
Ensure that you set
VLLM_SPYRE_USE_CB=1for decoding models.
12.8.1. IBM Spyre for Power AI acclerator card problems Copy linkLink copied to clipboard!
-
Ensure that the IBM Spyre AI accelerator cards are visible on the host. Use
lspcito verify that the cards are available. -
Ensure your user is in the
sentientgroup. - Use the Service Report tool to diagnose and correct card access issues. See IBM Power Systems service and productivity tools.
12.8.2. IBM Spyre for Power performance issues Copy linkLink copied to clipboard!
- Ensure all Spyre cards are securely seated in the first four slots of the IBM Power server I/O drawer. The first four slots have the highest speed PCIe interfaces.
- Ensure that cards assigned to an LPAR are all in the same drawer. Do not separate cards across drawers as this increases I/O latency. See IBM Power11 documentation for more information.
If you encounter errors with the IBM Spyre AI accelerator card, you can use the
aiu-smitool alongside the workload you want to profile. Perform the following steps:- Start the model.
From a second terminal, query the model. For example:
$ curl http://127.0.0.1:8000/v1/completions -H "Content-Type: application/json" \ -d '{ "model": "/models/granite-3.3-8b-instruct", "prompt": "Write me a long story about surfing dogs in Malibu.", "max_tokens": 8128, "temperature": 1, "n": 10 }'From a third terminal, run the
aiu-smitool:$ podman exec -it <CONTAINER_ID> -c aiu-smiAlternatively, exec into the running container and run
aiu-smi. For example:$ podman exec -it <CONTAINER_ID> bashRun the
aiu-smitool inside the container:[senuser@689230aca2ba ~]$ aiu-smiExample aiu-smi output
#MetricFiles # 0 /tmp/metrics.0181:50:00.0 # 1 /tmp/metrics.0182:60:00.0 # 2 /tmp/metrics.0183:70:00.0 # 3 /tmp/metrics.0184:80:00.0 #ID Date Time hostcpu hostmem pwr gtemp busy rdmem wrmem rxpci txpci rdrdma wrrdma rsvmem # YYYYMMDD HH:MM:SS % % W C % GB/s GB/s GB/s GB/s GB/s GB/s MB 0 20251103 20:18:36 951.6 11.5 33.8 34.1 96 41.221 5.480 0.967 0.964 0.000 0.000 0.000 1 20251103 20:18:36 951.6 11.5 30.6 33.0 96 41.201 5.464 0.967 0.964 0.000 0.000 0.000 2 20251103 20:18:36 951.6 11.5 40.5 34.7 96 41.266 5.473 0.969 0.966 0.000 0.000 0.000 3 20251103 20:18:36 951.6 11.5 37.3 39.2 96 41.358 5.484 0.971 0.968 0.000 0.000 0.000
Chapter 13. Gathering system information with the vLLM collect environment script Copy linkLink copied to clipboard!
Use the vllm collect-env command that you run from the Red Hat AI Inference Server container to gather system information for troubleshooting AI Inference Server deployments. This script collects system details, hardware configurations, and dependency information that can help diagnose deployment problems and model inference serving issues.
Prerequisites
- You have installed Podman or Docker.
- You are logged in as a user with sudo access.
- You have access to a Linux server with data center grade AI accelerators installed.
- You have pulled and successfully deployed the Red Hat AI Inference Server container.
Procedure
Open a terminal and log in to
registry.redhat.io:$ podman login registry.redhat.ioPull the specific Red Hat AI Inference Server container image for the AI accelerator that is installed. For example, to pull the Red Hat AI Inference Server container for Google cloud TPUs, run the following command:
$ podman pull registry.redhat.io/rhaii-early-access/vllm-tpu-rhel9:3.4.0-ea.1Run the collect environment script in the container:
$ podman run --rm -it \ --name vllm-tpu \ --network=host \ --privileged \ --device=/dev/vfio/vfio \ --device=/dev/vfio/0 \ -e PJRT_DEVICE=TPU \ -e HF_HUB_OFFLINE=0 \ -v ./.cache/rhaii:/opt/app-root/src/.cache:Z \ --entrypoint vllm collect-env \ registry.redhat.io/rhaii-early-access/vllm-tpu-rhel9:3.4.0-ea.1
Verification
The vllm collect-env command output details environment information including the following:
- System hardware details
- Operating system details
- Python environment and dependencies
- GPU/TPU accelerator information
Review the output for any warnings or errors that might indicate configuration issues. Include the collect-env output for your system when reporting problems to Red Hat Support.
An example Google Cloud TPU report is provided below:
==============================
System Info
==============================
OS : Red Hat Enterprise Linux 9.6 (Plow) (x86_64)
GCC version : (GCC) 11.5.0 20240719 (Red Hat 11.5.0-5)
Clang version : Could not collect
CMake version : version 4.1.0
Libc version : glibc-2.34
==============================
PyTorch Info
==============================
PyTorch version : 2.9.0.dev20250716
Is debug build : False
CUDA used to build PyTorch : None
ROCM used to build PyTorch : N/A
==============================
Python Environment
==============================
Python version : 3.12.9 (main, Jun 20 2025, 00:00:00) [GCC 11.5.0 20240719 (Red Hat 11.5.0-5)] (64-bit runtime)
Python platform : Linux-6.8.0-1015-gcp-x86_64-with-glibc2.34
==============================
CUDA / GPU Info
==============================
Is CUDA available : False
CUDA runtime version : No CUDA
CUDA_MODULE_LOADING set to : N/A
GPU models and configuration : No CUDA
Nvidia driver version : No CUDA
cuDNN version : No CUDA
HIP runtime version : N/A
MIOpen runtime version : N/A
Is XNNPACK available : True
==============================
CPU Info
==============================
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Address sizes: 52 bits physical, 57 bits virtual
Byte Order: Little Endian
CPU(s): 44
On-line CPU(s) list: 0-43
Vendor ID: AuthenticAMD
Model name: AMD EPYC 9B14
CPU family: 25
Model: 17
Thread(s) per core: 2
Core(s) per socket: 22
Socket(s): 1
Stepping: 1
BogoMIPS: 5200.00
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid tsc_known_freq pni pclmulqdq ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand hypervisor lahf_lm cmp_legacy cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw topoext ssbd ibrs ibpb stibp vmmcall fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves avx512_bf16 clzero xsaveerptr wbnoinvd arat avx512vbmi umip avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq rdpid fsrm
Hypervisor vendor: KVM
Virtualization type: full
L1d cache: 704 KiB (22 instances)
L1i cache: 704 KiB (22 instances)
L2 cache: 22 MiB (22 instances)
L3 cache: 96 MiB (3 instances)
NUMA node(s): 1
NUMA node0 CPU(s): 0-43
Vulnerability Gather data sampling: Not affected
Vulnerability Itlb multihit: Not affected
Vulnerability L1tf: Not affected
Vulnerability Mds: Not affected
Vulnerability Meltdown: Not affected
Vulnerability Mmio stale data: Not affected
Vulnerability Reg file data sampling: Not affected
Vulnerability Retbleed: Not affected
Vulnerability Spec rstack overflow: Mitigation; Safe RET
Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl
Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization
Vulnerability Spectre v2: Mitigation; Retpolines; IBPB conditional; IBRS_FW; STIBP always-on; RSB filling; PBRSB-eIBRS Not affected; BHI Not affected
Vulnerability Srbds: Not affected
Vulnerability Tsx async abort: Not affected
==============================
Versions of relevant libraries
==============================
[pip3] numpy==1.26.4
[pip3] pyzmq==27.0.1
[pip3] torch==2.9.0.dev20250716
[pip3] torch-xla==2.9.0.dev20250716
[pip3] torchvision==0.24.0.dev20250716
[pip3] transformers==4.55.2
[pip3] triton==3.3.1
[conda] Could not collect
==============================
vLLM Info
==============================
ROCM Version : Could not collect
Neuron SDK Version : N/A
vLLM Version : 0.10.0+rhai1
vLLM Build Flags:
CUDA Archs: Not Set; ROCm: Disabled; Neuron: Disabled
GPU Topology:
Could not collect
==============================
Environment Variables
==============================
VLLM_USE_V1=1
VLLM_WORKER_MULTIPROC_METHOD=spawn
VLLM_NO_USAGE_STATS=1
NCCL_CUMEM_ENABLE=0
PYTORCH_NVML_BASED_CUDA_CHECK=1
TORCHINDUCTOR_COMPILE_THREADS=1
TORCHINDUCTOR_CACHE_DIR=/tmp/torchinductor_default