이 콘텐츠는 선택한 언어로 제공되지 않습니다.
Preface
Red Hat AI Inference Server provides an OpenAI-compatible API server for inference serving. You can control the behavior of the server with arguments.
This document begins with a list of the most important server arguments that you use with the vllm serve
command. A complete list of vllm serve
arguments, environment variables, server metrics are also provided.