LocalRAG is a self-hosting application designed to deploy a production-level RAG chatbot on a local machine or server, ensuring data privacy and security. The application utilizes LangChain for backend processing, Streamlit for the frontend, Qdrant as the vector store, and Redis for storing chat messages. Prerequisites include setting up a Python virtual environment and Docker. Users also need to install Ollama for running large language models locally. The guide provides steps to clone the necessary repository and run the application using Streamlit.
Table of contents
Containerize and Deploy ML Models with FastAPI & Dockerawesome_llm_apps/LocalRAG at main · HemachandranD/awesome_llm_appsSort: