이 콘텐츠는 선택한 언어로 제공되지 않습니다.
LLM Compressor
Red Hat AI Inference Server 3.2
Compressing large language models with the LLM Compressor library
Abstract
Describes the LLM Compressor library and how you can use it to optimize and compress large language models before inferencing.