The AI agent world is messy. Teams are reaching for LangChain, LlamaIndex, CrewAI, AutoGen, or building custom solutions from scratch. Good. That's how it should be during the creative phase. But once an agent leaves a developer's laptop and starts talking to production data, calling external application programming interfaces (APIs), or running on shared infrastructure, freedom without guardrails stops being a feature and starts being a liability.
We've watched the industry go through waves: Model APIs (such as chat completions), agentic APIs (such as assistants and later the OpenAI responses API), the age of frameworks, and now the age of harnesses and coding agents. The top layer keeps changing. It's becoming fungible. What doesn't change is the gap between "it works on my laptop" and "it runs in production, securely, at scale, with audit trails."
Our AgentOps strategy is built on a core principle: Bring Your Own Agent (BYOA). The platform is framework-agnostic. What matters is that the agent has identity, runs under least-privilege, gets observed, passes safety checks, and can be audited after the fact. The platform provides security, governance, observability, and lifecycle management. The agent stays yours.
What Red Hat AI provides
This series previews how BYOA works in practice, covering what is available now and what we are building next.
We take OpenClaw, a personal AI assistant that routes agent interactions across channels (WhatsApp, Telegram, Slack, Discord, and more) through a central WebSocket Gateway, and we operationalize it on Red Hat AI. We aren’t wrapping it in a proprietary framework, we’re wrapping it in platform infrastructure. OpenClaw is just the example—this approach works for any agent runtime.
OpenClaw doesn't sandbox much by default. It doesn't enforce role-based access control (RBAC), trace tool calls, or gate access to external services. Red Hat AI adds each of those layers using Red Hat OpenShift and Red Hat OpenShift AI native capabilities, without touching the agent's code. The rest of the post walks through each of those layers.
Repurposing cloud-native to agent-native
We're not building agent infrastructure from scratch. We're repurposing the tools and deployment patterns that OpenShift already has and extending them for agentic workloads with Red Hat AI.
Agents need isolation: OpenShift sandboxed containers (based on the Kata Containers project, and a GA layered product) and the upcoming agent-sandbox integration provide kernel-isolated execution per agent session. The host and other agents' data are protected against a compromised agent.
Agents need identity: Scoped service-account tokens, with SPIFFE/SPIRE for cryptographic workload identity. No hardcoded keys. Instead, the platform will verify the agent and inject short‑lived, scoped service‑account tokens at the platform level. This is planned with the Kagenti integration for agent lifecycle as part of Red Hat AI.
Agents need multitenancy: Namespace isolation, NetworkPolicy, ResourceQuota, with verification that boundaries hold under adversarial testing.
Agents need policy guardrails at multiple layers: OPA/Gatekeeper and Kyverno policies at the Kubernetes level. The Model Context Protocol (MCP) Gateway for tool-level authorization. The Guardrails Orchestrator and NeMo Guardrails at the model inference boundary (more on that in the next section).
Making agents safer and ready for production
For many enterprise teams, safety is becoming a primary blocker for getting agents and models to production. Not performance. Not cost. Trust. Enterprises need confidence that their AI remains compliant with regulations and protects their brand from reputational and legal risks, especially once it's consumed by external users.
Red Hat AI provides a portfolio approach to safety that covers the full lifecycle, including detection, testing, and risk mitigation before they reach production, and guards against threats that emerge at runtime.
Before an agent goes live: Garak (planned as part of Red Hat AI) provides adversarial vulnerability scanning for jailbreaks, prompt injection, and other attack vectors at the model level. Integration is planned through the TrustyAI operator, and consumable via EvalHub (an evaluation control plane), and Kubeflow Pipelines, enabling adversarial scans in CI/CD before promotion.
At runtime (guardrails at the inference boundary): The TrustyAI Guardrails Orchestrator (generally available as of OpenShift AI 3.0) screens model inputs and outputs. NeMo Guardrails (tech preview) adds programmable conversational rails. Both operate at the inference boundary, intercepting individual large language model (LLM) calls to enforce safety before a response ever reaches the agent. A planned model-risk view in the model catalog will surface these safety signals alongside model metadata, so teams can factor in use-case risk before choosing a model.
Observability, tracing, and evaluation
Agents are stochastic. You cannot debug or trust them in production without execution traces.
Red Hat AI is providing the foundation for agent observability, starting with first-party support for MLflow, which is currently in developer preview, and planned for general availability in an upcoming release.
MLflow Tracing captures prompts, reasoning steps, tool invocations, LLM API requests, and token costs. All traces are OpenTelemetry compatible, so any OTEL-compatible sink can integrate. Beyond tracing, MLflow capabilities for scoring and evaluation can be used to assess agent quality over time, and these workflows are planned to be invoked through Eval Hub, an evaluation control plane that is part of OpenShift AI.
Governing tool calls at scale
Agents call tools. The question is how to govern those calls.
The MCP Gateway (built with the OpenShift networking team, Envoy-based), currently in developer preview, sits in front of all your MCP servers as a single, secure endpoint. It adds identity-based tool filtering so agents only see authorized tools, OAuth2 token exchange for scoped per-backend access, and credential management so sensitive tool calls go through proper authorization. The platform enforces access. The application manages credentials, so there is no cross-server leakage.
Authorization is enforced through Kuadrant's AuthPolicy, which integrates Authorino for JSON Web Token (JWT) validation and Open Policy Agent (OPA) rule evaluation at the Gateway API level.
For OpenClaw, this means the agent sets one MCP_URLenvironment variable and gets access to an aggregated tool catalog. Which tools it can actually call is determined by its token claims, not by the prompt. Prompt injection attacks that try to trick the agent into calling unauthorized tools get stopped at the infrastructure layer, because the gateway ignores prompt content entirely. It validates token claims.
Choosing API surfaces for production agents
A lot of teams started with chat, moved to chat completions and OpenAI APIs, then to frameworks, and now to agent harnesses. The APIs agents use are consolidating. One of the leading APIs for agentic workloads is the Responses API, and OpenAI has now opened that direction through the OpenResponses specification.
Red Hat AI is providing an implementation that is fully conformant with the OpenResponses specification. This creates a path for teams to run agent workloads on self-hosted or hybrid-model infrastructure, rather than routing every prompt, tool call, and reasoning artifact through third-party services.
OpenResponses-compatible runtimes for self-managed and hybrid environments are still limited. Red Hat AI provides one of the most mature implementations of the specification that targets that gap, making it a practical route for OpenClaw users who want to preserve OpenAI responses API-oriented agent behavior while moving execution to infrastructure they control.
For teams that want a self-hosted path without a Responses API orchestration layer, vLLM, part of Red Hat AI, provides an OpenAI-compatible/v1/chat/completionsendpoint that OpenClaw can consume directly.
Agent lifecycle with Kagenti
Many teams get stuck moving an agent from laptop to production. Kagenti, planned as part of OpenShift AI, bridges that gap. The kagenti-operator auto-discovers agents via A2A-based AgentCard CRDs and injects identity (SPIFFE/SPIRE), tracing, and tool governance (MCP Gateway) without changes to agent code via an AgentRuntime construct. The full lifecycle, from discovery to runtime governance, is managed by the platform. Agent catalog and registry are on the roadmap for the OpenShift AI UI, alongside an MCP catalog for tool servers
This series
This blog series walks you through these layers in detail with OpenClaw as an example agent runtime. Each post is self-contained. Read them in order or jump to whatever matches your current problem.
The only constant across every post is the BYOA principle. We never ask you to rewrite your agent. We bring enterprise rigor to the agent, not the other way around.
For more information about the Red Hat AI product stack for building, evaluating, and enforcing safety for AI applications, check the official product documentation as well as the following upstream projects that are key to the Red Hat AI AgentOps story:
리소스
적응형 엔터프라이즈: AI 준비성은 곧 위기 대응력
저자 소개
Adel Zaalouk is a product manager at Red Hat who enjoys blending business and technology to achieve meaningful outcomes. He has experience working in research and industry, and he's passionate about Red Hat OpenShift, cloud, AI and cloud-native technologies. He's interested in how businesses use OpenShift to solve problems, from helping them get started with containerization to scaling their applications to meet demand.
유사한 검색 결과
The efficient enterprise: Scaling intelligence with Mixture of Experts
Red Hat and NVIDIA collaborate for a more secure foundation for the agent-ready workforce
Technically Speaking | Build a production-ready AI toolbox
Technically Speaking | Platform engineering for AI agents
채널별 검색
오토메이션
기술, 팀, 인프라를 위한 IT 자동화 최신 동향
인공지능
고객이 어디서나 AI 워크로드를 실행할 수 있도록 지원하는 플랫폼 업데이트
오픈 하이브리드 클라우드
하이브리드 클라우드로 더욱 유연한 미래를 구축하는 방법을 알아보세요
보안
환경과 기술 전반에 걸쳐 리스크를 감소하는 방법에 대한 최신 정보
엣지 컴퓨팅
엣지에서의 운영을 단순화하는 플랫폼 업데이트
인프라
세계적으로 인정받은 기업용 Linux 플랫폼에 대한 최신 정보
애플리케이션
복잡한 애플리케이션에 대한 솔루션 더 보기
가상화
온프레미스와 클라우드 환경에서 워크로드를 유연하게 운영하기 위한 엔터프라이즈 가상화의 미래