Skip to content

Deployment of RAG + LLM model serving on multiple K8s cloud clusters

Notifications You must be signed in to change notification settings

elotl/GenAI-infra-stack

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Question-Answering in a Box with Self-hosted LLMs & RAG

  • Setup the complete infrastructure stack for a Question-Answering chatbot for your private data in just a few minutes!
  • Your stack will be powered by Self-hosted Open-Source Large Language Models and Retrieval Augmented Generation running on Kubernetes Cloud clusters.

Install Documentation

Jump to complete install doc available here.

About

Deployment of RAG + LLM model serving on multiple K8s cloud clusters

Topics

Resources

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Python 88.7%
  • Dockerfile 7.7%
  • Shell 3.6%