Universal LLM-Bridge für 600+ Modelle
LiteLLM ist die zentrale Inferenz-Schicht. Single API-Surface für Anthropic, OpenAI, OpenRouter, NVIDIA NIM, Ollama, Replicate, Together, Groq, Cerebras, Mistral, Cohere, Google, Bedrock, Azure, Fireworks, DeepSeek, Qwen. Cost-Routing: lokal-first, Failover-Chain, Per-User-Limits, Per-Team-Budgets. Audit-Logs in audit_events. Rate-Limiting, Caching, Streaming.