Skip to main content

The LiteLLM Blog

Guides, announcements, and best practices from the LiteLLM team.

Latest

Day 0 Support: GPT-5.4

GPT-5.4 model support in LiteLLM

openaigpt-5.4completion

DAY 0 Support: Gemini 3.1 Flash Lite Preview on LiteLLM

Guide to using Gemini 3.1 Flash Lite Preview on LiteLLM Proxy and SDK with day 0 support.

geminiday 0 supportllms

Incident Report: Cache Eviction Closes In-Use httpx Clients

Date: February 27, 2026

incident-reportcachingstability

Day 0 Support: GPT-5.3-Codex

Day 0 support for GPT-5.3-Codex on LiteLLM, including phase parameter handling for Responses API.

openaigpt-5.3-codexcodex

Incident Report: Encrypted Content Failures in Multi-Region Responses API Load Balancing

Date: Feb 24, 2026

incident-reportproxyresponses-api

Incident Report: Wildcard Blocking New Models After Cost Map Reload

Date: Feb 23, 2026

incident-reportproxyauth

Incident Report: SERVER_ROOT_PATH regression broke UI routing

Date: January 22, 2026

incident-reportuistability

DAY 0 Support: Gemini 3.1 Pro on LiteLLM

Guide to using Gemini 3.1 Pro on LiteLLM Proxy and SDK with day 0 support.

geminiday 0 supportllms

Incident Report: vLLM Embeddings Broken by encoding_format Parameter

Date: Feb 16, 2026

incident-reportembeddingsvllm

Day 0 Support: Claude Sonnet 4.6

Day 0 support for Claude Sonnet 4.6 on LiteLLM AI Gateway - use across Anthropic, Azure, Vertex AI, and Bedrock.

anthropicclaudesonnet 4.6