v1.63.11-stable
These are the changes since v1.63.2-stable.
This release is primarily focused on:
- [Beta] Responses API Support
 - Snowflake Cortex Support, Amazon Nova Image Generation
 - UI - Credential Management, re-use credentials when adding new models
 - UI - Test Connection to LLM Provider before adding a model
 
Known Issuesโ
- ๐จ Known issue on Azure OpenAI - We don't recommend upgrading if you use Azure OpenAI. This version failed our Azure OpenAI load test
 
Docker Run LiteLLM Proxyโ
docker run
-e STORE_MODEL_IN_DB=True
-p 4000:4000
ghcr.io/berriai/litellm:main-v1.63.11-stable
Demo Instanceโ
Here's a Demo Instance to test changes:
- Instance: https://demo.litellm.ai/
 - Login Credentials:
- Username: admin
 - Password: sk-1234
 
 
New Models / Updated Modelsโ
- Image Generation support for Amazon Nova Canvas Getting Started
 - Add pricing for Jamba new models PR
 - Add pricing for Amazon EU models PR
 - Add Bedrock Deepseek R1 model pricing PR
 - Update Gemini pricing: Gemma 3, Flash 2 thinking update, LearnLM PR
 - Mark Cohere Embedding 3 models as Multimodal PR
 - Add Azure Data Zone pricing PR
- LiteLLM Tracks cost for 
azure/euandazure/usmodels 
 - LiteLLM Tracks cost for 
 
LLM Translationโ
- New Endpoints
 
- [Beta] POST 
/responsesAPI. Getting Started 
- New LLM Providers
 
- Snowflake Cortex Getting Started
 
- New LLM Features
 
- Support OpenRouter 
reasoning_contenton streaming Getting Started 
- Bug Fixes
 
- OpenAI: Return 
code,paramandtypeon bad request error More information on litellm exceptions - Bedrock: Fix converse chunk parsing to only return empty dict on tool use PR
 - Bedrock: Support extra_headers PR
 - Azure: Fix Function Calling Bug & Update Default API Version to 
2025-02-01-previewPR - Azure: Fix AI services URL PR
 - Vertex AI: Handle HTTP 201 status code in response PR
 - Perplexity: Fix incorrect streaming response PR
 - Triton: Fix streaming completions bug PR
 - Deepgram: Support bytes.IO when handling audio files for transcription PR
 - Ollama: Fix "system" role has become unacceptable PR
 - All Providers (Streaming): Fix String 
data:stripped from entire content in streamed responses PR 
Spend Tracking Improvementsโ
- Support Bedrock converse cache token tracking Getting Started
 - Cost Tracking for Responses API Getting Started
 - Fix Azure Whisper cost tracking Getting Started
 
UIโ
Re-Use Credentials on UIโ
You can now onboard LLM provider credentials on LiteLLM UI. Once these credentials are added you can re-use them when adding new models Getting Started
Test Connections before adding modelsโ
Before adding a model you can test the connection to the LLM provider to verify you have setup your API Base + API Key correctly
General UI Improvementsโ
- Add Models Page
- Allow adding Cerebras, Sambanova, Perplexity, Fireworks, Openrouter, TogetherAI Models, Text-Completion OpenAI on Admin UI
 - Allow adding EU OpenAI models
 - Fix: Instantly show edit + deletes to models
 
 - Keys Page
- Fix: Instantly show newly created keys on Admin UI (don't require refresh)
 - Fix: Allow clicking into Top Keys when showing users Top API Key
 - Fix: Allow Filter Keys by Team Alias, Key Alias and Org
 - UI Improvements: Show 100 Keys Per Page, Use full height, increase width of key alias
 
 - Users Page
- Fix: Show correct count of internal user keys on Users Page
 - Fix: Metadata not updating in Team UI
 
 - Logs Page
- UI Improvements: Keep expanded log in focus on LiteLLM UI
 - UI Improvements: Minor improvements to logs page
 - Fix: Allow internal user to query their own logs
 - Allow switching off storing Error Logs in DB Getting Started
 
 - Sign In/Sign Out
- Fix: Correctly use 
PROXY_LOGOUT_URLwhen set Getting Started 
 - Fix: Correctly use 
 
Securityโ
- Support for Rotating Master Keys Getting Started
 - Fix: Internal User Viewer Permissions, don't allow 
internal_user_viewerrole to seeTest Key PageorCreate Key ButtonMore information on role based access controls - Emit audit logs on All user + model Create/Update/Delete endpoints Getting Started
 - JWT
- Support multiple JWT OIDC providers Getting Started
 - Fix JWT access with Groups not working when team is assigned All Proxy Models access
 
 - Using K/V pairs in 1 AWS Secret Getting Started
 
Logging Integrationsโ
- Prometheus: Track Azure LLM API latency metric Getting Started
 - Athina: Added tags, user_feedback and model_options to additional_keys which can be sent to Athina Getting Started
 
Performance / Reliability improvementsโ
- Redis + litellm router - Fix Redis cluster mode for litellm router PR
 
General Improvementsโ
- OpenWebUI Integration - display 
thinkingtokens 
- Guide on getting started with LiteLLM x OpenWebUI. Getting Started
 - Display 
thinkingtokens on OpenWebUI (Bedrock, Anthropic, Deepseek) Getting Started 
