Meta released Llama-3.3, and this post provides a hands-on demo for building a RAG app using it. The app allows users to interact with a document via chat. It uses LlamaIndex for orchestration, Qdrant for a self-hosted vector database, and Ollama for serving Llama-3.3 locally. The implementation steps include loading and
Sort: