Retrieval-augmented generation (RAG) gave AI a memory. Agents give it a job description. This captures where enterprise AI is today—the first wave focused on helping models say the right thing, but the next wave is about helping systems do the right thing.
Generative AI (gen AI) has moved quickly from experimentation with large language models (LLMs) to a race to operationalize AI at enterprise scale. For many organizations, RAG was the first practical step, grounding model outputs in enterprise data and making gen AI usable in real business contexts. But enterprises don't run on answers, they run on execution, and this is where RAG starts to show its limits.
RAG remains foundational
RAG solved a real and persistent problem. By connecting models to enterprise data at inference time, it helped make outputs more accurate, current, and defensible without the cost and complexity of retraining.
In doing this, RAG helped establish the patterns that made AI systems trustworthy. Data ingestion, preprocessing, and integration with internal knowledge sources help create a reliable baseline, and extensions like retrieval augmented fine-tuning further advanced that alignment by combining retrieval with model adaptation.
Where RAG breaks down
At its core, RAG follows a loop—retrieve, augment, respond. In practice, mature implementations go further, iterative retrieval, reranking, hypothetical document embeddings (HyDE), and hybrid search patterns all add meaningful complexity. But even the most sophisticated retrieval pipeline shares the same fundamental constraint—it's optimized to inform a response, not to carry out a workflow.
A single request might involve querying a customer relationship management (CRM) system, validating against internal systems, generating a response, routing the case, and logging the outcome for compliance. Retrieval is only one step in that much larger chain. To bridge the gap, teams assemble pipelines, connect APIs, and layer in orchestration logic. It works at first, but over time, those systems become fragile—they can be difficult to maintain, harder to govern, and increasingly opaque.
Agentic AI: from context to execution
Agentic AI changes the system’s role. Instead of responding to prompts, it works toward objectives, planning multistep workflows, making decisions based on intermediate results, calling tools and APIs, and adapting as conditions change. The model moves beyond generating responses to coordinating agent actions.
The foundation does not disappear, of course, retrieval and context still matter. What changes is how that context is used. Where RAG systems inform and contextualize, agentic systems execute and automate. And all of this is an extension of what came before. AI systems are shifting from responding to queries to achieving objectives, from generating text to producing outcomes, and from simple assistance to orchestrating work across entire systems.
Why openness and governance matter more
Once systems can autonomously execute actions, transparency and auditability become requirements. Organizations need to understand why a decision was made, what data it relied on, and how policies were enforced. Without that visibility, complexity turns into risk.
This is where architectural choices start to matter more. Systems that are inspectable, interoperable, and built on standard interfaces are easier to understand and govern. Systems that are closed or tightly coupled make those same questions harder to answer, so it is more difficult to operate safely at scale.
Bringing agents into production
Bringing agents into production is where the conversation changes. Up to this point, most teams have been able to move quickly by assembling frameworks and stitching together capabilities. Tools like LangChain, LangGraph, CrewAI, and AutoGen are capable frameworks that have proven their value, both in prototyping and in production. Many teams are already running meaningful workloads on them.
But as agentic systems grow in scope and start taking action across real enterprise environments, a different set of challenges emerges. Each integration is bespoke. Each workflow carries its own assumptions. Governance becomes something you retrofit rather than design for. The frameworks themselves are not the problem, the gap is in the foundation underneath them.
Agentic systems do not just need models and orchestration, they also need a consistent way to interact with tools, enforce policies, and operate across environments where data is already distributed. They need to be observable, governable, and designed as secure by default. This is the layer Red Hat AI is focused on.
What Red Hat AI provides
Red Hat AI isn't meant to be a replacement for the frameworks teams are already using, but to provide a more robust foundation that helps those systems run more reliably at scale, and be operated, governed, and trusted in production. The real challenge is making sure agents can continue to work safely and predictably as the environment around them evolves.
This becomes even more important in hybrid environments, where data, systems, and workloads are spread across clouds, data centers, and edge locations. In those contexts, centralization is not always practical, and forcing it often creates more problems than it solves. Systems have to move to where the data lives, and once systems can act autonomously, the cost of getting things wrong goes up fast. An incorrect response can be reviewed, but an incorrect action can execute immediately and cause irrevocable damage. This difference is what turns governance from a nice-to-have feature into an absolute requirement.
The path forward
The move from RAG to agentic AI is not a reset, it's a continuation. Organizations that invested in retrieval, data pipelines, and grounding models in real context already have the foundation they need and can effectively extend upon what they have already built. The question is whether those systems can evolve from answering questions to carrying out real work. The next phase of enterprise AI will not be defined by how well systems explain things, it will be defined by how well they operate. And in that, we believe, the model will matter less than the platform that supports it.
Try Red Hat OpenShift AI today
You can try OpenShift AI in the Red Hat Developer Sandbox with 30-day no-cost access to a fully managed environment. And be sure to register for Red Hat Summit 2026 to connect with our team and explore the future of production AI.
Risorsa
L'adattabilità enterprise: predisporsi all'IA per essere pronti a un'innovazione radicale
Sugli autori
As a principal technologist for AI at Red Hat with over 30 years of experience, Robbie works to support enterprise AI adoption through open source innovation. His focus is on cloud-native technologies, Kubernetes, and AI platforms, helping to deliver scalable and secure solutions using Red Hat AI.
Robbie is deeply committed to open source, open source AI, and open data, believing in the power of transparency, collaboration, and inclusivity to advance technology in meaningful ways. His work involves exploring private generative AI, traditional machine learning, and enhancing platform capabilities to support open and hybrid cloud solutions for AI. His focus is on helping organizations adopt ethical and sustainable AI technologies that make a real impact.
Frank La Vigne is a seasoned Data Scientist and the Principal Technical Marketing Manager for AI at Red Hat. He possesses an unwavering passion for harnessing the power of data to address pivotal challenges faced by individuals and organizations.
A trusted voice in the tech community, Frank co-hosts the renowned “Data Driven” podcast, a platform dedicated to exploring the dynamic domains of Data Science and Artificial Intelligence. Beyond his podcasting endeavors, he shares his insights and expertise through FranksWorld.com, a blog that serves as a testament to his dedication to the tech community. Always ahead of the curve, Frank engages with audiences through regular livestreams on LinkedIn, covering cutting-edge technological topics from quantum computing to the burgeoning metaverse.
Altri risultati simili a questo
The zero touch future: Enabling Telstra’s path to a fully autonomous, self-healing network
Integrating Red Hat Lightspeed with CrowdStrike for enhanced malware detection coverage
Technically Speaking | Build a production-ready AI toolbox
Technically Speaking | Platform engineering for AI agents
Ricerca per canale
Automazione
Novità sull'automazione IT di tecnologie, team e ambienti
Intelligenza artificiale
Aggiornamenti sulle piattaforme che consentono alle aziende di eseguire carichi di lavoro IA ovunque
Hybrid cloud open source
Scopri come affrontare il futuro in modo più agile grazie al cloud ibrido
Sicurezza
Le ultime novità sulle nostre soluzioni per ridurre i rischi nelle tecnologie e negli ambienti
Edge computing
Aggiornamenti sulle piattaforme che semplificano l'operatività edge
Infrastruttura
Le ultime novità sulla piattaforma Linux aziendale leader a livello mondiale
Applicazioni
Approfondimenti sulle nostre soluzioni alle sfide applicative più difficili
Virtualizzazione
Il futuro della virtualizzazione negli ambienti aziendali per i carichi di lavoro on premise o nel cloud