このコンテンツは選択した言語では利用できません。
Deploy models using Distributed Inference with llm-d
Red Hat OpenShift AI Self-Managed 3.4
Deploy and serve large language models at scale in Red Hat OpenShift AI
Abstract
As an administrator, you can use distributed inference to deploy and serve large language models at scale on Red Hat OpenShift AI.