An open-source Python SDK and proxy server for managing authentication, load balancing, and spend tracking across 100+ LLMs using a unified OpenAI format.
LiteLLM is an open-source toolkit that streamlines interactions with over 100 large language models (LLMs) by providing a unified API in the OpenAI format. It offers both a Python SDK and a proxy server (LLM Gateway) to manage authentication, load balancing, and spend tracking across various LLM providers, including OpenAI, Azure OpenAI, Vertex AI, and Amazon Bedrock. LiteLLM supports features such as retry and fallback logic, rate limiting, and logging integrations with tools like Langfuse and OpenTelemetry. It is designed to simplify the integration of multiple LLMs into applications, ensuring consistent output formats and efficient resource management.
We use cookies to enhance your experience. By continuing to use this site, you agree to our use of cookies. Learn more