Explore the advanced development of a conversational search agent using Ollama, Llama 3.1, Jina Embeddings, and ChromaDB. Learn how function calling enables the agent to interact with external web search tools for more accurate responses. Step-by-step instructions are provided for installing dependencies, creating a custom search tool, and setting up an agent system, complete with memory and tool execution capabilities.
Table of contents
Installing Ollama and Pulling ModelsInstalling and Importing DependenciesCreating a Custom Search ToolBuild The AgentInitializing the AgentRunning the Agent with Example QueriesAccessing the Agent’s MemorySort: