LiteLLM Gateway can now be deployed on Vercel, providing developers with an OpenAI-compatible interface to route LLM requests to any supported provider, including Vercel AI Gateway. A basic setup involves a Python entry point and a YAML config file to define model routing. A code snippet shows how to route a model through Vercel AI Gateway using the litellm_config.yaml file.
4 Comments
Sort: