このコンテンツは選択した言語では利用できません。
Preface
Distributed Inference with llm-d is a Kubernetes-native framework for serving large language models at scale. You can deploy Distributed Inference with llm-d on Openshift Container Platform or managed Kubernetes platforms such as Azure Kubernetes Service (AKS) and CoreWeave Kubernetes Service.
Distributed Inference with llm-d is a Technology Preview feature only. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process.
For more information about the support scope of Red Hat Technology Preview features, see Technology Preview Features Support Scope.