Learn how to test applications built on Large Language Models (LLMs) using Vercel's AI SDK and Vitest. This guide focuses on creating Evals—special tests for evaluating LLM performance—to ensure the Xata Agent works well after prompt modifications or model changes. The post details setting up the testing environment, organizing
Table of contents
The Xata AgentTesting the Agent with an EvalThe Eval run outputUsing Vitest to run an EvalConclusionSort: