Skip to main content

The LiteLLM Blog

Guides, announcements, and best practices from the LiteLLM team.

Latest

DAY 0 Support: Gemini 3.1 Pro on LiteLLM

Guide to using Gemini 3.1 Pro on LiteLLM Proxy and SDK with day 0 support.

geminiday 0 supportllms

Incident Report: vLLM Embeddings Broken by encoding_format Parameter

Date: Feb 16, 2026

incident-reportembeddingsvllm

Day 0 Support: Claude Sonnet 4.6

Day 0 support for Claude Sonnet 4.6 on LiteLLM AI Gateway - use across Anthropic, Azure, Vertex AI, and Bedrock.

anthropicclaudesonnet 4.6

Incident Report: Invalid beta headers with Claude Code

Date: February 13, 2026

incident-reportanthropicstability

Day 0 Support: MiniMax-M2.5

Day 0 support for MiniMax-M2.5 on LiteLLM

minimaxM2.5llm

Incident Report: Invalid model cost map on main

Date: January 27, 2026

incident-reportstability

Your Middleware Could Be a Bottleneck

How we improved LiteLLM proxy latency and throughput by replacing a single middleware base class

performancefastapimiddleware

Improve release stability with 24 hour load tests

How we built a long-running, release-validation system to catch regressions before they reach users.

testingobservabilityreliability

Day 0 Support: Claude Opus 4.6

Day 0 support for Claude Opus 4.6 on LiteLLM AI Gateway - use across Anthropic, Azure, Vertex AI, and Bedrock.

anthropicclaudeopus 4.6

Achieving Sub-Millisecond Proxy Overhead

Our Q1 performance target and architectural direction for achieving sub-millisecond proxy overhead on modest hardware.

performancearchitecture