Retrieval-augmented generation (RAG) gave AI a memory. Agents give it a job description. This captures where enterprise AI is today—the first wave focused on helping models say the right thing, but the next wave is about helping systems do the right thing.

Generative AI (gen AI) has moved quickly from experimentation with large language models (LLMs) to a race to operationalize AI at enterprise scale. For many organizations, RAG was the first practical step, grounding model outputs in enterprise data and making gen AI usable in real business contexts. But enterprises don't run on answers, they run on execution, and this is where RAG starts to show its limits.

RAG remains foundational

RAG solved a real and persistent problem. By connecting models to enterprise data at inference time, it helped make outputs more accurate, current, and defensible without the cost and complexity of retraining.

In doing this, RAG helped establish the patterns that made AI systems trustworthy. Data ingestion, preprocessing, and integration with internal knowledge sources help create a reliable baseline, and extensions like retrieval augmented fine-tuning further advanced that alignment by combining retrieval with model adaptation.

Where RAG breaks down

At its core, RAG follows a loop—retrieve, augment, respond. In practice, mature implementations go further, iterative retrieval, reranking, hypothetical document embeddings (HyDE), and hybrid search patterns all add meaningful complexity. But even the most sophisticated retrieval pipeline shares the same fundamental constraint—it's optimized to inform a response, not to carry out a workflow. 

A single request might involve querying a customer relationship management (CRM) system, validating against internal systems, generating a response, routing the case, and logging the outcome for compliance. Retrieval is only one step in that much larger chain. To bridge the gap, teams assemble pipelines, connect APIs, and layer in orchestration logic. It works at first, but over time, those systems become fragile—they can be difficult to maintain, harder to govern, and increasingly opaque.

Agentic AI: from context to execution

Agentic AI changes the system’s role. Instead of responding to prompts, it works toward objectives, planning multistep workflows, making decisions based on intermediate results, calling tools and APIs, and adapting as conditions change. The model moves beyond generating responses to coordinating agent actions.

The foundation does not disappear, of course, retrieval and context still matter. What changes is how that context is used. Where RAG systems inform and contextualize, agentic systems execute and automate. And all of this is an extension of what came before. AI systems are shifting from responding to queries to achieving objectives, from generating text to producing outcomes, and from simple assistance to orchestrating work across entire systems.

Why openness and governance matter more

Once systems can autonomously execute actions, transparency and auditability become requirements. Organizations need to understand why a decision was made, what data it relied on, and how policies were enforced. Without that visibility, complexity turns into risk.

This is where architectural choices start to matter more. Systems that are inspectable, interoperable, and built on standard interfaces are easier to understand and govern. Systems that are closed or tightly coupled make those same questions harder to answer, so it is more difficult to operate safely at scale.

Bringing agents into production

Bringing agents into production is where the conversation changes. Up to this point, most teams have been able to move quickly by assembling frameworks and stitching together capabilities. Tools like LangChain, LangGraph, CrewAI, and AutoGen are capable frameworks that have proven their value, both in prototyping and in production. Many teams are already running meaningful workloads on them.

But as agentic systems grow in scope and start taking action across real enterprise environments, a different set of challenges emerges. Each integration is bespoke. Each workflow carries its own assumptions. Governance becomes something you retrofit rather than design for. The frameworks themselves are not the problem, the gap is in the foundation underneath them.

Agentic systems do not just need models and orchestration, they also need a consistent way to interact with tools, enforce policies, and operate across environments where data is already distributed. They need to be observable, governable, and designed as secure by default. This is the layer Red Hat AI is focused on.

What Red Hat AI provides

Red Hat AI isn't meant to be a replacement for the frameworks teams are already using, but to provide a more robust foundation that helps those systems run more reliably at scale, and  be operated, governed, and trusted in production. The real challenge is making sure agents can continue to work safely and predictably as the environment around them evolves.

This becomes even more important in hybrid environments, where data, systems, and workloads are spread across clouds, data centers, and edge locations. In those contexts, centralization is not always practical, and forcing it often creates more problems than it solves. Systems have to move to where the data lives, and once systems can act autonomously, the cost of getting things wrong goes up fast. An incorrect response can be reviewed, but an incorrect action can execute immediately and cause irrevocable damage. This difference is what turns governance from a nice-to-have feature into an absolute requirement.

The path forward

The move from RAG to agentic AI is not a reset, it's a continuation. Organizations that invested in retrieval, data pipelines, and grounding models in real context already have the foundation they need and can effectively extend upon what they have already built. The question is whether those systems can evolve from answering questions to carrying out real work. The next phase of enterprise AI will not be defined by how well systems explain things, it will be defined by how well they operate. And in that, we believe, the model will matter less than the platform that supports it.

Try Red Hat OpenShift AI today

You can try OpenShift AI in the Red Hat Developer Sandbox with 30-day no-cost access to a fully managed environment. And be sure to register for Red Hat Summit 2026 to connect with our team and explore the future of production AI.

资源

自适应企业:AI 就绪,从容应对颠覆性挑战

这本由红帽首席运营官兼首席战略官 Michael Ferris 撰写的电子书,介绍了当今 IT 领导者面临的 AI 变革和技术颠覆挑战。

关于作者

As a principal technologist for AI at Red Hat with over 30 years of experience, Robbie works to support enterprise AI adoption through open source innovation. His focus is on cloud-native technologies, Kubernetes, and AI platforms, helping to deliver scalable and secure solutions using Red Hat AI.

Robbie is deeply committed to open source, open source AI, and open data, believing in the power of transparency, collaboration, and inclusivity to advance technology in meaningful ways. His work involves exploring private generative AI, traditional machine learning, and enhancing platform capabilities to support open and hybrid cloud solutions for AI. His focus is on helping organizations adopt ethical and sustainable AI technologies that make a real impact.

Frank La Vigne is a seasoned Data Scientist and the Principal Technical Marketing Manager for AI at Red Hat. He possesses an unwavering passion for harnessing the power of data to address pivotal challenges faced by individuals and organizations.
A trusted voice in the tech community, Frank co-hosts the renowned “Data Driven” podcast, a platform dedicated to exploring the dynamic domains of Data Science and Artificial Intelligence. Beyond his podcasting endeavors, he shares his insights and expertise through FranksWorld.com, a blog that serves as a testament to his dedication to the tech community. Always ahead of the curve, Frank engages with audiences through regular livestreams on LinkedIn, covering cutting-edge technological topics from quantum computing to the burgeoning metaverse.

UI_Icon-Red_Hat-Close-A-Black-RGB

按频道浏览

automation icon

自动化

有关技术、团队和环境 IT 自动化的最新信息

AI icon

人工智能

平台更新使客户可以在任何地方运行人工智能工作负载

open hybrid cloud icon

开放混合云

了解我们如何利用混合云构建更灵活的未来

security icon

安全防护

有关我们如何跨环境和技术减少风险的最新信息

edge icon

边缘计算

简化边缘运维的平台更新

Infrastructure icon

基础架构

全球领先企业 Linux 平台的最新动态

application development icon

应用领域

我们针对最严峻的应用挑战的解决方案

Virtualization icon

虚拟化

适用于您的本地或跨云工作负载的企业虚拟化的未来