Skip to content

Commit d808211

Browse files
authored
Merge pull request #567 from jeremyary/chatqna-amd-rh
Add docs to accompany AMD-accelerated OPEA ChatQnA pattern
2 parents 1722ede + b68e5bb commit d808211

17 files changed

+517
-0
lines changed
Lines changed: 111 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,111 @@
1+
---
2+
title: OPEA Chat QnA accelerated with AMD Instinct
3+
date: 2025-05-01
4+
tier: sandbox
5+
validated: false
6+
summary: This pattern aids with deployment of OPEA's Chat QnA RAG application, accelerated with AMD Instinct hardware.
7+
rh_products:
8+
- Red Hat OpenShift Container Platform
9+
- Red Hat OpenShift Data Foundation
10+
- Red Hat OpenShift AI
11+
- Red Hat OpenShift Serverless
12+
- Red Hat OpenShift Service Mesh
13+
partners:
14+
- AMD
15+
industries:
16+
- General
17+
aliases: /amd-rag-chat-qna/
18+
#pattern_logo: amd-rag-chat-qna.png
19+
links:
20+
install: amd-rag-chat-qna-getting-started
21+
help: https://groups.google.com/g/validatedpatterns
22+
bugs: https://github.com/validatedpatterns-sandbox/
23+
---
24+
:toc:
25+
:imagesdir: /images
26+
:_content-type: ASSEMBLY
27+
include::modules/comm-attributes.adoc[]
28+
29+
30+
[id="about-amd-rag-chat-qna-pattern"]
31+
= About {amdqna-pattern}
32+
33+
Background::
34+
This Validated Pattern implements an enterprise-ready question-and-answer chatbot utilizing retrieval-augmented generation (RAG) and capability reasoning using large language model (LLM). The application is based on the publicly-available https://github.com/opea-project/GenAIExamples/tree/main/ChatQnA[OPEA Chat QnA] example application created by the https://opea.dev/[Open Platform for Enterprise AI (OPEA)] community.
35+
36+
OPEA provides a framework that enables the creation and evaluation of open, multi-provider, robust, and composable generative AI (GenAI) solutions. It harnesses the best innovations across the ecosystem while keeping enterprise-level needs front and center. It simplifies the implementation of enterprise-grade composite GenAI solutions, starting with a focus on Retrieval Augmented Generative AI (RAG). The platform is designed to facilitate efficient integration of secure, performant, and cost-effective GenAI workflows into business systems and manage its deployments, leading to quicker GenAI adoption and business value.
37+
38+
This pattern aims to leverage the strengths of OPEA's framework in combination with other OpenShift-centric toolings in order to deploy a modern, LLM-backed reasoning application stack capable of leveraging https://www.amd.com/en/products/accelerators/instinct.html[AMD's Instinct] hardware acceleration in an enterprise-ready and distributed manner. The pattern utilizes https://www.redhat.com/en/technologies/cloud-computing/openshift/gitops[Red Hat OpenShift GitOps] to bring a continuous delivery approach to the application's development and usage based on declarative Git-driven workflows, backed by a centralized, single-source-of-truth git repository.
39+
40+
Key features::
41+
- AMD Instinct GPU acceleration for high-performance AI inferencing
42+
- GitOps-based deployment and management through Red Hat Validated Patterns
43+
- OPEA-based AI/ML pipeline with specialized services for document processing
44+
- Enterprise-grade security with HashiCorp Vault integration
45+
- Vector database support for efficient similarity search and retrieval
46+
47+
[id="about-solution"]
48+
== About the solution
49+
50+
The following solution integrates the https://github.com/opea-project/GenAIExamples/tree/main/ChatQnA[OPEA ChatQnA example app] with the http://localhost:1313/patterns/multicloud-gitops/[Multicloud GitOps] Validated Pattern to encapsulate every defined component as an easily trackable resource via OpenShift GitOps dashboard:
51+
52+
Components::
53+
* AI/ML Services
54+
** Text Embeddings Inference (TEI)
55+
** Document Retriever
56+
** Retriever Service
57+
** LLM-TGI (Text Generation Inference) from OPEA
58+
** vLLM accelerated by AMD Instinct GPUs
59+
** Redis Vector Database
60+
* Infrastructure
61+
** Red Hat OpenShift AI (RHOAI)
62+
** AMD GPU Operator
63+
** OpenShift Data Foundation (ODF)
64+
** Kernel Module Management (KMM)
65+
** Node Feature Discovery (NFD)
66+
* Security
67+
** HashiCorp Vault
68+
** External Secrets Operator
69+
70+
.Overview of the solution
71+
image::/images/qna-chat-amd/amd-rag-chat-qna-overview.png[alt=OPEA Chat QnA accelerated with AMD Instinct Validated Pattern architecture,65%,65%]
72+
73+
.Overview of application flow
74+
image::/images/qna-chat-amd/amd-rag-chat-qna-flow.png[OPEA Chat QnA accelerated with AMD Instinct application flow]
75+
76+
[id="about-technology"]
77+
== About the technology
78+
79+
The following technologies are used in this solution:
80+
81+
https://www.redhat.com/en/technologies/cloud-computing/openshift/try-it[Red Hat OpenShift Platform]::
82+
An enterprise-ready Kubernetes container platform built for an open hybrid cloud strategy. It provides a consistent application platform to manage hybrid cloud, public cloud, and edge deployments. It delivers a complete application platform for both traditional and cloud-native applications, allowing them to run anywhere. OpenShift has a pre-configured, pre-installed, and self-updating monitoring stack that provides monitoring for core platform components. It also enables the use of external secret management systems, for example, HashiCorp Vault in this case, to securely add secrets into the OpenShift platform.
83+
84+
https://www.redhat.com/en/technologies/cloud-computing/openshift/try-it[Red Hat OpenShift GitOps]::
85+
A declarative application continuous delivery tool for Kubernetes based on the ArgoCD project. Application definitions, configurations, and environments are declarative and version controlled in Git. It can automatically push the desired application state into a cluster, quickly find out if the application state is in sync with the desired state, and manage applications in multi-cluster environments.
86+
87+
https://www.redhat.com/en/technologies/management/ansible[Red Hat Ansible Automation Platform]::
88+
Provides an enterprise framework for building and operating IT automation at scale across hybrid clouds including edge deployments. It enables users across an organization to create, share, and manage automation, from development and operations to security and network teams.
89+
90+
https://www.redhat.com/en/technologies/cloud-computing/openshift-data-foundation[Red Hat OpenShift Data Foundation]::
91+
It is software-defined storage for containers. Red Hat OpenShift Data Foundation helps teams develop and deploy applications quickly and efficiently across clouds.
92+
93+
https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai[Red Hat OpenShift AI]::
94+
Red Hat® OpenShift® AI is a flexible, scalable artificial intelligence (AI) and machine learning (ML) platform that enables enterprises to create and deliver AI-enabled applications at scale across hybrid cloud environments.
95+
96+
https://www.redhat.com/en/technologies/cloud-computing/openshift/serverless[Red Hat OpenShift Serverless]::
97+
Red Hat® OpenShift® Serverless simplifies the development of hybrid cloud applications by eliminating the complexities associated with Kubernetes and the infrastructure applications are developed and deployed on. Developers will be able to focus on coding applications instead of managing intricate infrastructure details.
98+
99+
https://www.redhat.com/en/technologies/cloud-computing/openshift/what-is-openshift-service-mesh[Red Hat OpenShift Service Mesh]::
100+
Red Hat® OpenShift® Service Mesh provides a uniform way to connect, manage, and observe microservices-based applications. It provides behavioral insight into—and control of—the networked microservices in your service mesh.
101+
102+
https://catalog.redhat.com/software/container-stacks/detail/61954b7020da7eae27db0e2a[Hashicorp Vault]::
103+
Provides a secure centralized store for dynamic infrastructure and applications across clusters, including over low-trust networks between clouds and data centers.
104+
105+
https://github.com/opea-project[OPEA]::
106+
OPEA is an ecosystem orchestration framework to integrate performant GenAI technologies and workflows leading to quicker GenAI adoption and business value.
107+
108+
[id="next-steps_mcg-index"]
109+
== Next steps
110+
111+
* link:amd-rag-chat-qna-getting-started[Deploy the pattern].

0 commit comments

Comments
 (0)