Retrieval-augmented generation (RAG) gave AI a memory. Agents give it a job description. This captures where enterprise AI is today—the first wave focused on helping models say the right thing, but the next wave is about helping systems do the right thing.

Generative AI (gen AI) has moved quickly from experimentation with large language models (LLMs) to a race to operationalize AI at enterprise scale. For many organizations, RAG was the first practical step, grounding model outputs in enterprise data and making gen AI usable in real business contexts. But enterprises don't run on answers, they run on execution, and this is where RAG starts to show its limits.

RAG remains foundational

RAG solved a real and persistent problem. By connecting models to enterprise data at inference time, it helped make outputs more accurate, current, and defensible without the cost and complexity of retraining.

In doing this, RAG helped establish the patterns that made AI systems trustworthy. Data ingestion, preprocessing, and integration with internal knowledge sources help create a reliable baseline, and extensions like retrieval augmented fine-tuning further advanced that alignment by combining retrieval with model adaptation.

Where RAG breaks down

At its core, RAG follows a loop—retrieve, augment, respond. In practice, mature implementations go further, iterative retrieval, reranking, hypothetical document embeddings (HyDE), and hybrid search patterns all add meaningful complexity. But even the most sophisticated retrieval pipeline shares the same fundamental constraint—it's optimized to inform a response, not to carry out a workflow. 

A single request might involve querying a customer relationship management (CRM) system, validating against internal systems, generating a response, routing the case, and logging the outcome for compliance. Retrieval is only one step in that much larger chain. To bridge the gap, teams assemble pipelines, connect APIs, and layer in orchestration logic. It works at first, but over time, those systems become fragile—they can be difficult to maintain, harder to govern, and increasingly opaque.

Agentic AI: from context to execution

Agentic AI changes the system’s role. Instead of responding to prompts, it works toward objectives, planning multistep workflows, making decisions based on intermediate results, calling tools and APIs, and adapting as conditions change. The model moves beyond generating responses to coordinating agent actions.

The foundation does not disappear, of course, retrieval and context still matter. What changes is how that context is used. Where RAG systems inform and contextualize, agentic systems execute and automate. And all of this is an extension of what came before. AI systems are shifting from responding to queries to achieving objectives, from generating text to producing outcomes, and from simple assistance to orchestrating work across entire systems.

Why openness and governance matter more

Once systems can autonomously execute actions, transparency and auditability become requirements. Organizations need to understand why a decision was made, what data it relied on, and how policies were enforced. Without that visibility, complexity turns into risk.

This is where architectural choices start to matter more. Systems that are inspectable, interoperable, and built on standard interfaces are easier to understand and govern. Systems that are closed or tightly coupled make those same questions harder to answer, so it is more difficult to operate safely at scale.

Bringing agents into production

Bringing agents into production is where the conversation changes. Up to this point, most teams have been able to move quickly by assembling frameworks and stitching together capabilities. Tools like LangChain, LangGraph, CrewAI, and AutoGen are capable frameworks that have proven their value, both in prototyping and in production. Many teams are already running meaningful workloads on them.

But as agentic systems grow in scope and start taking action across real enterprise environments, a different set of challenges emerges. Each integration is bespoke. Each workflow carries its own assumptions. Governance becomes something you retrofit rather than design for. The frameworks themselves are not the problem, the gap is in the foundation underneath them.

Agentic systems do not just need models and orchestration, they also need a consistent way to interact with tools, enforce policies, and operate across environments where data is already distributed. They need to be observable, governable, and designed as secure by default. This is the layer Red Hat AI is focused on.

What Red Hat AI provides

Red Hat AI isn't meant to be a replacement for the frameworks teams are already using, but to provide a more robust foundation that helps those systems run more reliably at scale, and  be operated, governed, and trusted in production. The real challenge is making sure agents can continue to work safely and predictably as the environment around them evolves.

This becomes even more important in hybrid environments, where data, systems, and workloads are spread across clouds, data centers, and edge locations. In those contexts, centralization is not always practical, and forcing it often creates more problems than it solves. Systems have to move to where the data lives, and once systems can act autonomously, the cost of getting things wrong goes up fast. An incorrect response can be reviewed, but an incorrect action can execute immediately and cause irrevocable damage. This difference is what turns governance from a nice-to-have feature into an absolute requirement.

The path forward

The move from RAG to agentic AI is not a reset, it's a continuation. Organizations that invested in retrieval, data pipelines, and grounding models in real context already have the foundation they need and can effectively extend upon what they have already built. The question is whether those systems can evolve from answering questions to carrying out real work. The next phase of enterprise AI will not be defined by how well systems explain things, it will be defined by how well they operate. And in that, we believe, the model will matter less than the platform that supports it.

Try Red Hat OpenShift AI today

You can try OpenShift AI in the Red Hat Developer Sandbox with 30-day no-cost access to a fully managed environment. And be sure to register for Red Hat Summit 2026 to connect with our team and explore the future of production AI.

Ressource

L'entreprise adaptable : quand s'adapter à l'IA signifie s'adapter aux changements

Ce livre numérique de Michael Ferris, directeur de l'exploitation et de la stratégie chez Red Hat, aborde le rythme des changements et des bouleversements technologiques liés à l'IA auxquels sont confrontés les responsables informatiques.

À propos des auteurs

As a principal technologist for AI at Red Hat with over 30 years of experience, Robbie works to support enterprise AI adoption through open source innovation. His focus is on cloud-native technologies, Kubernetes, and AI platforms, helping to deliver scalable and secure solutions using Red Hat AI.

Robbie is deeply committed to open source, open source AI, and open data, believing in the power of transparency, collaboration, and inclusivity to advance technology in meaningful ways. His work involves exploring private generative AI, traditional machine learning, and enhancing platform capabilities to support open and hybrid cloud solutions for AI. His focus is on helping organizations adopt ethical and sustainable AI technologies that make a real impact.

Frank La Vigne is a seasoned Data Scientist and the Principal Technical Marketing Manager for AI at Red Hat. He possesses an unwavering passion for harnessing the power of data to address pivotal challenges faced by individuals and organizations.
A trusted voice in the tech community, Frank co-hosts the renowned “Data Driven” podcast, a platform dedicated to exploring the dynamic domains of Data Science and Artificial Intelligence. Beyond his podcasting endeavors, he shares his insights and expertise through FranksWorld.com, a blog that serves as a testament to his dedication to the tech community. Always ahead of the curve, Frank engages with audiences through regular livestreams on LinkedIn, covering cutting-edge technological topics from quantum computing to the burgeoning metaverse.

UI_Icon-Red_Hat-Close-A-Black-RGB

Parcourir par canal

automation icon

Automatisation

Les dernières nouveautés en matière d'automatisation informatique pour les technologies, les équipes et les environnements

AI icon

Intelligence artificielle

Actualité sur les plateformes qui permettent aux clients d'exécuter des charges de travail d'IA sur tout type d'environnement

open hybrid cloud icon

Cloud hybride ouvert

Découvrez comment créer un avenir flexible grâce au cloud hybride

security icon

Sécurité

Les dernières actualités sur la façon dont nous réduisons les risques dans tous les environnements et technologies

edge icon

Edge computing

Actualité sur les plateformes qui simplifient les opérations en périphérie

Infrastructure icon

Infrastructure

Les dernières nouveautés sur la plateforme Linux d'entreprise leader au monde

application development icon

Applications

À l’intérieur de nos solutions aux défis d’application les plus difficiles

Virtualization icon

Virtualisation

L'avenir de la virtualisation d'entreprise pour vos charges de travail sur site ou sur le cloud