LiteLLM logo

LiteLLM

LiteLLM AI Agent
Rating:
Rate it!

Overview

An open-source Python SDK and proxy server for managing authentication, load balancing, and spend tracking across 100+ LLMs using a unified OpenAI format.

LiteLLM is an open-source toolkit that streamlines interactions with over 100 large language models (LLMs) by providing a unified API in the OpenAI format. It offers both a Python SDK and a proxy server (LLM Gateway) to manage authentication, load balancing, and spend tracking across various LLM providers, including OpenAI, Azure OpenAI, Vertex AI, and Amazon Bedrock. LiteLLM supports features such as retry and fallback logic, rate limiting, and logging integrations with tools like Langfuse and OpenTelemetry. It is designed to simplify the integration of multiple LLMs into applications, ensuring consistent output formats and efficient resource management.

Some of the use cases of LiteLLM:

  • Developers seeking a unified interface to interact with multiple LLM providers.
  • Implementing load balancing and fallback mechanisms across different LLMs.
  • Tracking and managing API usage and spending across various projects.
  • Integrating LLM functionalities into applications with consistent output formats.

LiteLLM Video:

We use cookies to enhance your experience. By continuing to use this site, you agree to our use of cookies. Learn more