We are excited to share that KServe, the leading standardized AI inference platform on Kubernetes, has been accepted as an incubating project by the Cloud Native Computing Foundation (CNCF).

This milestone validates KServe’s maturity, stability and role as the foundation for scalable, multi-framework model serving in production environments. By moving into the CNCF’s neutral governance, KServe’s development will be driven purely by community needs, accelerating its standardization for serving AI models on Kubernetes.

For Red Hat this is a validation of our commitment to delivering open, reliable and standardized AI solutions for the hybrid cloud.

The critical engine behind Red Hat OpenShift AI

At Red Hat, we believe the best AI infrastructure is built on open standards and Kubernetes. KServe is the critical model serving component that powers Red Hat OpenShift AI, helping ensure our customers can transition from model experimentation to production inference seamlessly and at scale.

OpenShift AI leverages KServe’s features to solve the biggest enterprise AI challenges, helping enterprises realize:

  • High-performance LLM optimization - KServe is optimized for large language models (LLMs), providing high-performance features like KV cache offloading, distributed inference with vLLM, as well as disaggregated serving, pre-fix caching, intelligent scheduling and variant autoscaling via the integration with llm-d.
  • Advanced autoscaling - In addition to the horizontal pod autoscaling capability from Kubernetes, KServe also supports autoscaling with KEDA (Kubernetes Event-driven Autoscaler), which enables event-driven scaling based on external metrics such as vLLM metrics.
  • Both predictive and generative AI model inference - KServe supports pluggable, reusable, extensible runtimes, ranging from scikit-learn and XGBoost for predictive AI to Hugging Face and vLLM for generative AI model inference. This helps ensure that enterprises can switch to the best runtime for specific use cases.

Unlocking enterprise AI value

The journey of AI from the lab to the bottom line requires production infrastructure that can handle exponential growth, especially as enterprise usage shifts to widespread generative applications.

Now bolstered by the full resources and neutral governance of the CNCF, KServe directly addresses these core operational challenges - from tackling complexity with a unified API to controlling cloud costs through its scale-to-zero capabilities.

This move offers enterprises confidence in the longevity, security and open future of their AI infrastructure investment. KServe is now positioned to be the open-source standard for cloud-native model serving, empowering enterprises to confidently build, deploy and scale the next generation of intelligent applications on a foundation of open standards and community-driven innovation.

Join the Movement!

We invite the community to join us in congratulating the KServe maintainers and contributors on this achievement.

Want to get involved and shape the future of AI inference? Here’s how you can join the journey:

Check out KServe GitHub repo: https://github.com/kserve/kserve
Join KServe community: https://github.com/kserve/community 

Additionally, join us at our KubeCon+CloudNativeCon NA sessions to hear more about KServe:

Anchoring Trust in the Age of AI: Identities Across Humans, Machines, and Models - Tuesday November 11, 2025 10:08 am ET


关于作者

Yuan is a Senior Principal Software Engineer at Red Hat AI. Previously, he has led AI infrastructure and platform teams at various companies. He holds leadership positions in open source communities, including Argo, Kubeflow, KServe, Kubernetes, and CNCF. He's also a maintainer and author of many popular open source projects. In addition, Yuan authored three technical books as well as numerous papers and patents. He's a frequent conference speaker, technical advisor, leader, and mentor at various organizations.

UI_Icon-Red_Hat-Close-A-Black-RGB

按频道浏览

automation icon

自动化

有关技术、团队和环境 IT 自动化的最新信息

AI icon

人工智能

平台更新使客户可以在任何地方运行人工智能工作负载

open hybrid cloud icon

开放混合云

了解我们如何利用混合云构建更灵活的未来

security icon

安全防护

有关我们如何跨环境和技术减少风险的最新信息

edge icon

边缘计算

简化边缘运维的平台更新

Infrastructure icon

基础架构

全球领先企业 Linux 平台的最新动态

application development icon

应用领域

我们针对最严峻的应用挑战的解决方案

Virtualization icon

虚拟化

适用于您的本地或跨云工作负载的企业虚拟化的未来