Get started with AI Inference: Red Hat AI experts explain

This e-book introduces the fundamentals of inference performance engineering and model optimization, with a focus on quantization, sparsity, and other techniques that help reduce compute and memory requirements for artificial intelligence (AI) models. It highlights the benefits of using a Red Hat® open approach, validated model repository, and tools like the LLM Compressor and Red Hat AI Inference Server. Download to get started.

 

Front cover of Get started with AI Inference ebook featuring a large stylized 3D star