Welcome to our second edition of the monthly vLLM roundup! We are excited to continue sharing updates about the project, new features, and opportunities to engage with the vLLM community. Check out the December roundup here.
Keep on reading for exciting updates. And please share this post to others who may benefit!
Upcoming bi-weekly vLLM Office Hours
Distributed inference with vLLM | January 23, 2025 - 2:00PM ET / 11:00AM PT
Join our upcoming vLLM Office Hours as we dive into distributed inference with vLLM. We'll explore common pitfalls, practical implementation strategies, and steps to get started, with insights tailored to real-world challenges like those discussed here.
Recent recordings
vLLM’s 2024 wrapped and 2025 vision
vLLM v0.6.6 Update & open discussion
Blog posts
Structured decoding in vLLM: A gentle introduction
vLLM is the high-throughput and efficient inference engine for running large-language models (LLMs). In this post, we will explore the annotated history of language models, describe the current state of structured decoding in vLLM, as well as the recent integration with XGrammar, and share our tentative roadmap for future improvements.
vLLM 2024 retrospective and 2025 vision
The vLLM community achieved remarkable growth in 2024, evolving from a specialized inference engine to becoming the de facto serving solution for the open-source AI ecosystem. Celebrate vLLMs 2024 achievements and get a sneak peek into the 2025 roadmap.
Installing and Developing vLLM with Ease
The field of LLM inference is advancing at an unprecedented pace. With new models and features emerging weekly, the traditional software release pipeline often struggles to keep up. With vLLM, we aim to provide more than just a software package. We are building a dynamic ecosystem that adapts to this rapid evolution, offering developers the tools, documentation, and community support they need to stay ahead.
2:4 Sparse Llama FP8: SOTA Performance for NVIDIA Hopper GPUs
Advancing AI efficiency is more critical than ever, and sparsity has proven to be a cornerstone in this pursuit. Building on our previous work at Neural Magic with the 2:4 Sparse Llama 3.1 8B foundation model–which increases model efficiency by eliminating unnecessary parameters while preserving accuracy–we are excited to introduce the next step forward: Sparse 8-bit floating point (FP8) models and the associated high-performance kernels for vLLM.
Events
1️⃣ The year of full-stack OSS AI!
Optimizing LLMs for Cost-Efficient Deployment with vLLM
Michael Goin, Neural Magic [Red Hat]
Deploying LLMs is just the starting point; optimizing them for cost-efficient, high-performance serving is the real challenge. In this talk, we’ll explore cutting-edge compression techniques and advanced inference system optimizations that enable fast performance on your hardware of choice. Discover practical strategies and tools enterprises trust to scale deployments while minimizing costs.
2️⃣ West coast vLLM meetup
The first vLLM meetup in 2025 is on Wednesday, January 22nd in San Francisco. We will discuss vLLM's performant V1 architecture, Q1 roadmap, and Google Cloud's innovation around vLLM: networking, Cloud Run, Vertex, and TPU!
3️⃣ First-ever east coast vLLM meetup
It’s happening on March 11, 2025, in Boston! More details coming in early February.
In other news
It’s official! Red Hat completed the acquisition of Neural Magic! By acquiring Neural Magic, a leading commercial contributor to vLLM, Red Hat aims to continue supporting the vibrant vLLM community and enhancing Red Hat AI’s ability to support gen AI deployments anywhere and everywhere across the hybrid cloud. Read more on the completed acquisition here.
vLLM is nearing 34,000 stars! 🌟 Be sure to add your star and join the community. Thank you for your support.
Recurso
Introducción a la inteligencia artificial para las empresas: Guía para principiantes
Sobre el autor
Saša Zelenović is a Principal Product Marketing Manager at Red Hat, joining in 2025 through the Neural Magic acquisition where he led as Head of Marketing. With a passion for developer-focused marketing, Sasa drives efforts to help developers compress models for inference and deploy them with vLLM. He co-hosts the bi-weekly vLLM Office Hours, a go-to spot for insights and community around all things vLLM.
Más como éste
How Red Hat OpenShift AI simplifies trust and compliance
A 5-step playbook for unified automation and AI
Technically Speaking | Platform engineering for AI agents
Technically Speaking | Driving healthcare discoveries with AI
Navegar por canal
Automatización
Las últimas novedades en la automatización de la TI para los equipos, la tecnología y los entornos
Inteligencia artificial
Descubra las actualizaciones en las plataformas que permiten a los clientes ejecutar cargas de trabajo de inteligecia artificial en cualquier lugar
Nube híbrida abierta
Vea como construimos un futuro flexible con la nube híbrida
Seguridad
Vea las últimas novedades sobre cómo reducimos los riesgos en entornos y tecnologías
Edge computing
Conozca las actualizaciones en las plataformas que simplifican las operaciones en el edge
Infraestructura
Vea las últimas novedades sobre la plataforma Linux empresarial líder en el mundo
Aplicaciones
Conozca nuestras soluciones para abordar los desafíos más complejos de las aplicaciones
Virtualización
El futuro de la virtualización empresarial para tus cargas de trabajo locales o en la nube