Many organizations are considering generative AI (gen AI) as a means of reducing the time employees spend filing IT requests and the time IT teams spend closing these requests. To this end, we’ve created an it-self-service-agent AI quickstart that shows you how to automate IT processes within your organization using agentic AI on Red Hat OpenShift AI

This AI quickstart provides a reusable framework—including request routing, agent services, knowledge bases, an integration dispatcher, and an evaluation framework—that you can apply across multiple IT processes. While we're demonstrating a laptop refresh process as the working example, the same components can be adapted to privacy impact assessments, RFP generation, access requests, software licensing, and other structured IT workflows.

AI quickstarts are a catalog of ready-to-run, industry-specific use cases for your Red Hat AI environment. Each AI quickstart is designed to be simple to deploy, explore, and extend, giving teams a fast, hands-on way to see how AI can power solutions on enterprise-ready, open source infrastructure. You can read more about AI quickstarts in the article, “AI quickstarts: An easy and practical way to get started with Red Hat AI.”

The business case

Using AI to automate IT processes has a variety of benefits, including:

  • Reducing time needed to complete common requests: The agent helps employees create their requests by explaining the options and required information, and helps employees submit those requests once they are ready.
  • Improving compliance to process standards: Requests are more complete and aligned with process standards. This reduces the need to contact the requesting employee for additional information and reduces the amount of time and effort needed to review and complete requests.
  • Reducing the number of rejected requests due to missing/incorrect information: Rejected requests are frustrating for employees and lead to lower employee satisfaction. Reducing request rejection and eliminating the back and forth on requests helps improve employee satisfaction.
  • Reducing ticket resolution times: The agent helps reduce the time needed to close a ticket, improving throughput and reducing ticket idle time.

Meeting employees where they are

One of the key aspects of realizing these benefits is the design philosophy of, "meeting employees where they are." Rather than requiring employees to learn a new system or communication channel, this demo integrates with the tools your organization already uses. If your organization uses Slack and email to communicate with employees, integrating with these existing channels can improve adoption rates and reduce change management overhead compared to introducing yet another new tool. Employees can interact with AI agents using the same familiar interfaces they use for all their other work communication.

To that end, this AI quickstart includes production-ready integrations with Slack, email, and ServiceNow, along with a general-purpose request manager that makes integrating with other existing communication channels straightforward. The top-level architecture is as follows:

A diagram of an AI agent processing user input from several channels and accessing data sources and APIs

Figure 1: Architecture diagram showing the IT self-service agent components including request manager, agent services, knowledge bases, integration dispatcher, and connections to Slack, email, and ServiceNow

Getting started

Time to complete: 60-90 minutes (depending on whether you explore optional integrations like Slack and ServiceNow).

This AI quickstart is designed for rapid exploration and hands-on learning. In about an hour, you can deploy a fully functional AI agent system and see it handle complete laptop refresh workflows. The optional integrations with Slack, ServiceNow, and email can be added incrementally as you explore further.

Once you have the prerequisites in place, you can install the AI quickstart as shown below. The prerequisites include an existing Red Hat OpenShift environment and a number of local tools as outlined in the requirements.

# Clone the repository
git clone https://github.com/rh-ai-quickstart/it-self-service-agent.git
cd it-self-service-agent

# Set your namespace
export NAMESPACE=your-namespace

# Set LLM configuration
export LLM=llama-3-3-70b-instruct-w8a8
export LLM_ID=llama-3-3-70b-instruct-w8a8
export LLM_API_TOKEN=your-api-token
export LLM_URL=https://your-llm-endpoint

# Log in to OpenShift
oc login --server=https://your-cluster:6443

# Deploy in testing mode (Mock Eventing)
make helm-install-test NAMESPACE=$NAMESPACE

This will deploy the set of pods that make up the quickstart:

A topology view with nine blue circular icons, each with the Helm logo and a corresponding label

Figure 2: Screenshot showing the deployed OpenShift pods that make up the IT self-service agent AI quickstart including agent services, request manager, integration dispatcher, and supporting infrastructure

Once you have the prerequisite OpenShift pods in place, you are now ready to explore the key aspects of the agentic implementation of a laptop refresh process. The AI quickstart takes you through:

  • Requesting a laptop refresh through a command-line interface.
  • Integration with Slack and requesting a laptop refresh through Slack.
  • Integration with a real ServiceNow instance and viewing the request you created in the ServiceNow UI.
  • Enabling email support and requesting a laptop refresh through email.
  • Running multi-turn evaluations on the conversations with the laptop refresh agents using the open source DeepEval framework. Gen AI agents are non-deterministic, so traditional software testing is insufficient. This evaluation framework validates business-specific requirements—such as policy compliance and information gathering—so agents meet quality standards before deployment, and catch regressions during updates. This is critical for confidently deploying and iterating on AI agents in production.
  • Viewing details of the requests to and responses from the laptop refresh agents using OpenShift's built-in OpenTelemetry support, including calls through the Responses API and calls to MCP servers. Agentic AI systems involve complex interactions between multiple components, making production debugging challenging without proper visibility. Distributed tracing enables teams to understand how requests flow through the system, identify performance bottlenecks, diagnose issues in production, and understand user interaction patterns—essential capabilities for maintaining reliable AI agent systems at scale.
  • Exploring different prompting approaches and the trade-offs between a single large prompt and multistep prompts using LangGraph.
  • Setting up PromptGuard to avoid prompt injection attacks.
  • Setting up LlamaGuard to add content moderation.

All of this runs on Red Hat OpenShift AI with production-level scalability built in. By completing this AI quickstart, you will have:

  • Deployed a fully functional AI agent system on Red Hat OpenShift AI
  • Understood the core platform architecture and components
  • Tested the laptop refresh agent through multiple channels
  • Run evaluations to validate agent behavior
  • Implemented safety checks with PromptGuard and LlamaGuard
  • Explored AI observability with OpenTelemetry
  • Learned how to customize the system for your own use cases

To learn more

This AI quickstart helps you explore each of the key aspects of automating an IT process with agents and includes a deep dive on each of the components. 

We hope you're excited to try out this IT self-service agent AI quickstart. If you find it interesting, be sure to check out the full AI quickstarts catalog—there you'll find other use cases that will help you and your team start using Red Hat AI to solve real-world problems and to inspire you to think about what else you might build with the platform.

리소스

적응형 엔터프라이즈: AI 준비성은 곧 위기 대응력

Red Hat의 COO 겸 CSO인 Michael Ferris가 쓴 이 e-Book은 오늘날 IT 리더들이 직면한 AI의 변화와 기술적 위기의 속도를 살펴봅니다.

저자 소개

Michael Dawson is an active contributor to the Node.js project and chair of the Node.js Technical Steering Committee (TSC). He contributes to a broad range of community efforts including platform support, build infrastructure, N-API, Release, as well as tools to help the community achieve quality with speed (e.g., ci jobs, benchmarking and code coverage reporting). As the Node.js lead for Red Hat and IBM, he works with Red Hat and IBM's internal teams to plan and facilitate their contributions to Node.js and v8 within the Node and Google communities. Dawson's past experience includes building IBM's Java runtime, building and operating client facing e-commerce applications, building PKI and symmetric based crypto solutions as well as a number of varied consulting engagements. In his spare time, he uses Node.js to automate his home and life for fun.

UI_Icon-Red_Hat-Close-A-Black-RGB

채널별 검색

automation icon

오토메이션

기술, 팀, 인프라를 위한 IT 자동화 최신 동향

AI icon

인공지능

고객이 어디서나 AI 워크로드를 실행할 수 있도록 지원하는 플랫폼 업데이트

open hybrid cloud icon

오픈 하이브리드 클라우드

하이브리드 클라우드로 더욱 유연한 미래를 구축하는 방법을 알아보세요

security icon

보안

환경과 기술 전반에 걸쳐 리스크를 감소하는 방법에 대한 최신 정보

edge icon

엣지 컴퓨팅

엣지에서의 운영을 단순화하는 플랫폼 업데이트

Infrastructure icon

인프라

세계적으로 인정받은 기업용 Linux 플랫폼에 대한 최신 정보

application development icon

애플리케이션

복잡한 애플리케이션에 대한 솔루션 더 보기

Virtualization icon

가상화

온프레미스와 클라우드 환경에서 워크로드를 유연하게 운영하기 위한 엔터프라이즈 가상화의 미래