このコンテンツは選択した言語では利用できません。
Red Hat AI Model Optimization Toolkit
Red Hat AI Inference 3.4
Compressing large language models with the LLM Compressor library
Abstract
Describes the LLM Compressor library and how you can use it to optimize and compress large language models before inferencing.