<?xml version="1.0" encoding="utf-8"?>
<feed xmlns="http://www.w3.org/2005/Atom">
    <id>https://docs.litellm.ai/release_notes</id>
    <title>liteLLM Blog</title>
    <updated>2026-02-28T00:00:00.000Z</updated>
    <generator>https://github.com/jpmonette/feed</generator>
    <link rel="alternate" href="https://docs.litellm.ai/release_notes"/>
    <subtitle>liteLLM Blog</subtitle>
    <icon>https://docs.litellm.ai/img/favicon.ico</icon>
    <entry>
        <title type="html"><![CDATA[v1.82.0 - Realtime Guardrails, Projects Management, and 10+ Performance Optimizations]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-82-0</id>
        <link href="https://docs.litellm.ai/release_notes/v1-82-0"/>
        <updated>2026-02-28T00:00:00.000Z</updated>
        <summary type="html"><![CDATA[Deploy this version]]></summary>
        <content type="html"><![CDATA[<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-82-0#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<!-- -->
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">ghcr.io/berriai/litellm:main-1.82.0-stable</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.82.0</span><br></span></code></pre></div></div></div></div></div>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-82-0#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<ul>
<li><strong>Realtime API guardrails</strong> — <a href="https://docs.litellm.ai/docs/proxy/guardrails">Full guardrails support for <code>/v1/realtime</code> WebSocket sessions with pre/post-call enforcement, voice transcription hooks, session termination policies, and Vertex AI Gemini Live support</a> - <a href="https://github.com/BerriAI/litellm/pull/22152" target="_blank" rel="noopener noreferrer">PR #22152</a>, <a href="https://github.com/BerriAI/litellm/pull/22153" target="_blank" rel="noopener noreferrer">PR #22153</a>, <a href="https://github.com/BerriAI/litellm/pull/22161" target="_blank" rel="noopener noreferrer">PR #22161</a>, <a href="https://github.com/BerriAI/litellm/pull/22165" target="_blank" rel="noopener noreferrer">PR #22165</a></li>
<li><strong>Projects Management</strong> — <a href="https://docs.litellm.ai/docs/proxy/ui_store_model_db_setting">New Projects UI with full CRUD, project-scoped virtual keys, and admin opt-in toggle — organize teams and keys by project</a> - <a href="https://github.com/BerriAI/litellm/pull/22315" target="_blank" rel="noopener noreferrer">PR #22315</a>, <a href="https://github.com/BerriAI/litellm/pull/22360" target="_blank" rel="noopener noreferrer">PR #22360</a>, <a href="https://github.com/BerriAI/litellm/pull/22373" target="_blank" rel="noopener noreferrer">PR #22373</a>, <a href="https://github.com/BerriAI/litellm/pull/22412" target="_blank" rel="noopener noreferrer">PR #22412</a></li>
<li><strong>Guardrail ecosystem expansion</strong> — <a href="https://docs.litellm.ai/docs/proxy/guardrails">Noma v2, Lakera v2 post-call, Singapore regulatory policies (PDPA + MAS), employment discrimination blockers, code execution blocker, guardrail policy versioning, and production monitoring</a> - <a href="https://github.com/BerriAI/litellm/pull/21400" target="_blank" rel="noopener noreferrer">PR #21400</a>, <a href="https://github.com/BerriAI/litellm/pull/21783" target="_blank" rel="noopener noreferrer">PR #21783</a>, <a href="https://github.com/BerriAI/litellm/pull/21948" target="_blank" rel="noopener noreferrer">PR #21948</a></li>
<li><strong>OpenAI Codex 5.3 — day 0</strong> — <a href="https://docs.litellm.ai/docs/providers/openai">Full support for <code>gpt-5.3-codex</code> on OpenAI and Azure, plus <code>gpt-audio-1.5</code> and <code>gpt-realtime-1.5</code> model coverage</a> - <a href="https://github.com/BerriAI/litellm/pull/22035" target="_blank" rel="noopener noreferrer">PR #22035</a></li>
<li><strong>10+ performance optimizations</strong> — Streaming hot-path fixes, Redis pipeline batching, database task batching, ModelResponse init skip, and router cache improvements — lower latency and CPU on every request</li>
<li><strong><code>/v1/messages</code> → <code>/responses</code> routing</strong> — <code>/v1/messages</code> requests are now routed to the <a href="https://docs.litellm.ai/docs/response_api">Responses API</a> by default for OpenAI/Azure models</li>
</ul>
<div class="theme-admonition theme-admonition-danger admonition_xJq3 alert alert--danger"><div class="admonitionHeading_Gvgb"><span class="admonitionIcon_Rf37"><svg viewBox="0 0 12 16"><path fill-rule="evenodd" d="M5.05.31c.81 2.17.41 3.38-.52 4.31C3.55 5.67 1.98 6.45.9 7.98c-1.45 2.05-1.7 6.53 3.53 7.7-2.2-1.16-2.67-4.52-.3-6.61-.61 2.03.53 3.33 1.94 2.86 1.39-.47 2.3.53 2.27 1.67-.02.78-.31 1.44-1.13 1.81 3.42-.59 4.78-3.42 4.78-5.56 0-2.84-2.53-3.22-1.25-5.61-1.52.13-2.03 1.13-1.89 2.75.09 1.08-1.02 1.8-1.86 1.33-.67-.41-.66-1.19-.06-1.78C8.18 5.31 8.68 2.45 5.05.32L5.03.3l.02.01z"></path></svg></span>v1/messages routing change</div><div class="admonitionContent_BuS1"><p>This version starts routing <code>/v1/messages</code> requests to the <code>/responses</code> API by default. To opt out and continue using chat/completions, set <code>LITELLM_USE_CHAT_COMPLETIONS_URL_FOR_ANTHROPIC_MESSAGES=true</code> or <code>litellm_settings.use_chat_completions_url_for_anthropic_messages: true</code> in your config.</p></div></div>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-82-0#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support-20-new-models">New Model Support (20 new models)<a href="https://docs.litellm.ai/release_notes/v1-82-0#new-model-support-20-new-models" class="hash-link" aria-label="Direct link to New Model Support (20 new models)" title="Direct link to New Model Support (20 new models)">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th><th>Features</th></tr></thead><tbody><tr><td>OpenAI</td><td><code>gpt-5.3-codex</code></td><td>272K</td><td>$1.75</td><td>$14.00</td><td>Reasoning, coding</td></tr><tr><td>Azure OpenAI</td><td><code>azure/gpt-5.3-codex</code></td><td>272K</td><td>$1.75</td><td>$14.00</td><td>Azure deployment</td></tr><tr><td>OpenAI</td><td><code>gpt-audio-1.5</code></td><td>128K</td><td>$2.50</td><td>$10.00</td><td>Audio model</td></tr><tr><td>Azure OpenAI</td><td><code>azure/gpt-audio-1.5-2026-02-23</code></td><td>128K</td><td>$2.50</td><td>$10.00</td><td>Audio model</td></tr><tr><td>OpenAI</td><td><code>gpt-realtime-1.5</code></td><td>32K</td><td>$4.00</td><td>$16.00</td><td>Realtime model</td></tr><tr><td>Azure OpenAI</td><td><code>azure/gpt-realtime-1.5-2026-02-23</code></td><td>32K</td><td>$4.00</td><td>$16.00</td><td>Realtime model</td></tr><tr><td>Groq</td><td><code>groq/openai/gpt-oss-safeguard-20b</code></td><td>131K</td><td>$0.075</td><td>$0.30</td><td>Guardrail inference</td></tr><tr><td>Google Vertex AI</td><td><code>vertex_ai/gemini-3.1-flash-image-preview</code></td><td>-</td><td>-</td><td>-</td><td>Image generation</td></tr><tr><td>Perplexity</td><td><code>perplexity/perplexity/sonar</code></td><td>-</td><td>-</td><td>-</td><td>Sonar search</td></tr><tr><td>Perplexity</td><td><code>perplexity/openai/gpt-5.1</code></td><td>-</td><td>-</td><td>-</td><td>Hosted routing</td></tr><tr><td>Perplexity</td><td><code>perplexity/openai/gpt-5-mini</code></td><td>-</td><td>-</td><td>-</td><td>Hosted routing</td></tr><tr><td>Perplexity</td><td><code>perplexity/google/gemini-2.5-flash</code></td><td>-</td><td>-</td><td>-</td><td>Hosted routing</td></tr><tr><td>Perplexity</td><td><code>perplexity/google/gemini-2.5-pro</code></td><td>-</td><td>-</td><td>-</td><td>Hosted routing</td></tr><tr><td>Perplexity</td><td><code>perplexity/google/gemini-3-flash-preview</code></td><td>-</td><td>-</td><td>-</td><td>Hosted routing</td></tr><tr><td>Perplexity</td><td><code>perplexity/google/gemini-3-pro-preview</code></td><td>-</td><td>-</td><td>-</td><td>Hosted routing</td></tr><tr><td>Perplexity</td><td><code>perplexity/anthropic/claude-haiku-4-5</code></td><td>-</td><td>-</td><td>-</td><td>Hosted routing</td></tr><tr><td>Perplexity</td><td><code>perplexity/anthropic/claude-sonnet-4-5</code></td><td>-</td><td>-</td><td>-</td><td>Hosted routing</td></tr><tr><td>Perplexity</td><td><code>perplexity/anthropic/claude-opus-4-5</code></td><td>-</td><td>-</td><td>-</td><td>Hosted routing</td></tr><tr><td>Perplexity</td><td><code>perplexity/anthropic/claude-opus-4-6</code></td><td>-</td><td>-</td><td>-</td><td>Hosted routing</td></tr><tr><td>Perplexity</td><td><code>perplexity/xai/grok-4-1-fast-non-reasoning</code></td><td>-</td><td>-</td><td>-</td><td>Hosted routing</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-82-0#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI</a></strong></p>
<ul>
<li>Day 0 support for <code>gpt-5.3-codex</code> on OpenAI and Azure - <a href="https://github.com/BerriAI/litellm/pull/22035" target="_blank" rel="noopener noreferrer">PR #22035</a></li>
<li>Add <code>gpt-audio-1.5</code> model cost map - <a href="https://github.com/BerriAI/litellm/pull/22303" target="_blank" rel="noopener noreferrer">PR #22303</a></li>
<li>Add <code>gpt-realtime-1.5</code> model cost map - <a href="https://github.com/BerriAI/litellm/pull/22304" target="_blank" rel="noopener noreferrer">PR #22304</a></li>
<li>Add <code>audio</code> as supported OpenAI param - <a href="https://github.com/BerriAI/litellm/pull/22092" target="_blank" rel="noopener noreferrer">PR #22092</a></li>
<li>Add <code>prompt_cache_key</code> and <code>prompt_cache_retention</code> support - <a href="https://github.com/BerriAI/litellm/pull/20397" target="_blank" rel="noopener noreferrer">PR #20397</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/azure">Azure OpenAI</a></strong></p>
<ul>
<li>New Azure OpenAI models 2026-02-25 - <a href="https://github.com/BerriAI/litellm/pull/22114" target="_blank" rel="noopener noreferrer">PR #22114</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Add v1 Anthropic Responses API transformation - <a href="https://github.com/BerriAI/litellm/pull/22087" target="_blank" rel="noopener noreferrer">PR #22087</a></li>
<li>Sanitize <code>tool_use</code> IDs in <code>convert_to_anthropic_tool_invoke</code> - <a href="https://github.com/BerriAI/litellm/pull/21964" target="_blank" rel="noopener noreferrer">PR #21964</a></li>
<li>Fix model wildcard access issue - <a href="https://github.com/BerriAI/litellm/pull/21917" target="_blank" rel="noopener noreferrer">PR #21917</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">AWS Bedrock</a></strong></p>
<ul>
<li>Encode model ARNs for OpenAI-compatible Bedrock imported models - <a href="https://github.com/BerriAI/litellm/pull/21701" target="_blank" rel="noopener noreferrer">PR #21701</a></li>
<li>Support optional regional STS endpoint in role assumption - <a href="https://github.com/BerriAI/litellm/pull/21640" target="_blank" rel="noopener noreferrer">PR #21640</a></li>
<li>Native structured outputs API support - <a href="https://github.com/BerriAI/litellm/pull/21222" target="_blank" rel="noopener noreferrer">PR #21222</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vertex">Google Vertex AI</a></strong></p>
<ul>
<li>Add <code>gemini-3.1-flash-image-preview</code> to model cost map - <a href="https://github.com/BerriAI/litellm/pull/22223" target="_blank" rel="noopener noreferrer">PR #22223</a></li>
<li>Enable <code>context-1m-2025-08-07</code> beta header for Vertex AI provider - <a href="https://github.com/BerriAI/litellm/pull/21867" target="_blank" rel="noopener noreferrer">PR #21867</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openrouter">OpenRouter</a></strong></p>
<ul>
<li>Add OpenRouter native models to model cost map - <a href="https://github.com/BerriAI/litellm/pull/20520" target="_blank" rel="noopener noreferrer">PR #20520</a></li>
<li>Add OpenRouter Opus 4.6 to model map - <a href="https://github.com/BerriAI/litellm/pull/20525" target="_blank" rel="noopener noreferrer">PR #20525</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/mistral">Mistral</a></strong></p>
<ul>
<li>Adjust <code>mistral-small-2503</code> input/output cost per token - <a href="https://github.com/BerriAI/litellm/pull/22097" target="_blank" rel="noopener noreferrer">PR #22097</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/groq">Groq</a></strong></p>
<ul>
<li>Add <code>groq/openai/gpt-oss-safeguard-20b</code> model pricing - <a href="https://github.com/BerriAI/litellm/pull/21951" target="_blank" rel="noopener noreferrer">PR #21951</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/aiml">AI/ML</a></strong></p>
<ul>
<li>Update AIML model pricing - <a href="https://github.com/BerriAI/litellm/pull/22139" target="_blank" rel="noopener noreferrer">PR #22139</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/ollama">Ollama</a></strong></p>
<ul>
<li>Thread <code>api_base</code> to <code>get_model_info</code> + graceful fallback - <a href="https://github.com/BerriAI/litellm/pull/21970" target="_blank" rel="noopener noreferrer">PR #21970</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openai">PublicAI</a></strong></p>
<ul>
<li>Fix function calling for PublicAI Apertus models - <a href="https://github.com/BerriAI/litellm/pull/21582" target="_blank" rel="noopener noreferrer">PR #21582</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/xai">xAI</a></strong></p>
<ul>
<li>Add deprecation dates for <code>grok-2-vision-1212</code> and <code>grok-3-mini</code> models - <a href="https://github.com/BerriAI/litellm/pull/20102" target="_blank" rel="noopener noreferrer">PR #20102</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Forward auth headers of provider - <a href="https://github.com/BerriAI/litellm/pull/22070" target="_blank" rel="noopener noreferrer">PR #22070</a></li>
<li>Normalize camelCase <code>thinking</code> param keys to snake_case - <a href="https://github.com/BerriAI/litellm/pull/21762" target="_blank" rel="noopener noreferrer">PR #21762</a></li>
<li>Allow <code>dimensions</code> param passthrough for non-text-embedding-3 OpenAI models - <a href="https://github.com/BerriAI/litellm/pull/22144" target="_blank" rel="noopener noreferrer">PR #22144</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-82-0#bug-fixes" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">AWS Bedrock</a></strong></p>
<ul>
<li>Fix converse handling for <code>parallel_tool_calls</code> - <a href="https://github.com/BerriAI/litellm/pull/22267" target="_blank" rel="noopener noreferrer">PR #22267</a></li>
<li>Restore <code>parallel_tool_calls</code> mapping in <code>map_openai_params</code> - <a href="https://github.com/BerriAI/litellm/pull/22333" target="_blank" rel="noopener noreferrer">PR #22333</a></li>
<li>Correct <code>modelInput</code> format for Converse API batch models - <a href="https://github.com/BerriAI/litellm/pull/21656" target="_blank" rel="noopener noreferrer">PR #21656</a></li>
<li>Prevent double UUID in <code>create_file</code> S3 key - <a href="https://github.com/BerriAI/litellm/pull/21650" target="_blank" rel="noopener noreferrer">PR #21650</a></li>
<li>Filter internal <code>json_tool_call</code> when mixed with real tools - <a href="https://github.com/BerriAI/litellm/pull/21107" target="_blank" rel="noopener noreferrer">PR #21107</a></li>
<li>Pass timeout param to Bedrock rerank HTTP client - <a href="https://github.com/BerriAI/litellm/pull/22021" target="_blank" rel="noopener noreferrer">PR #22021</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Fix model cost map for anthropic fast and <code>inference_geo</code> - <a href="https://github.com/BerriAI/litellm/pull/21904" target="_blank" rel="noopener noreferrer">PR #21904</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/image_generation">Image Generation</a></strong></p>
<ul>
<li>Propagate <code>extra_headers</code> to upstream image generation - <a href="https://github.com/BerriAI/litellm/pull/22026" target="_blank" rel="noopener noreferrer">PR #22026</a></li>
<li>Add <code>ChatCompletionImageObject</code> in <code>OpenAIChatCompletionAssistantMessage</code> - <a href="https://github.com/BerriAI/litellm/pull/22155" target="_blank" rel="noopener noreferrer">PR #22155</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Preserve forwarding of server-side called tools - <a href="https://github.com/BerriAI/litellm/pull/22260" target="_blank" rel="noopener noreferrer">PR #22260</a></li>
<li>Fix free model handling from UI paths - <a href="https://github.com/BerriAI/litellm/pull/22258" target="_blank" rel="noopener noreferrer">PR #22258</a></li>
<li>Fix <code>None</code> TypeError in mapping - <a href="https://github.com/BerriAI/litellm/pull/22080" target="_blank" rel="noopener noreferrer">PR #22080</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-82-0#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-82-0#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/response_api">Realtime API</a></strong></p>
<ul>
<li>Guardrails support for <code>/v1/realtime</code> WebSocket endpoint - <a href="https://github.com/BerriAI/litellm/pull/22152" target="_blank" rel="noopener noreferrer">PR #22152</a></li>
<li>Vertex AI Gemini Live via unified <code>/realtime</code> endpoint - <a href="https://github.com/BerriAI/litellm/pull/22153" target="_blank" rel="noopener noreferrer">PR #22153</a></li>
<li>Guardrails with <code>pre_call</code>/<code>post_call</code> mode on realtime WebSocket - <a href="https://github.com/BerriAI/litellm/pull/22161" target="_blank" rel="noopener noreferrer">PR #22161</a></li>
<li><code>end_session_after_n_fails</code> + Endpoint Settings wizard step - <a href="https://github.com/BerriAI/litellm/pull/22165" target="_blank" rel="noopener noreferrer">PR #22165</a></li>
<li>Guardrail hook for voice transcription - <a href="https://github.com/BerriAI/litellm/pull/21976" target="_blank" rel="noopener noreferrer">PR #21976</a></li>
<li>Fix guardrails not firing for Gemini/Vertex AI and <code>provider_config</code> realtime sessions - <a href="https://github.com/BerriAI/litellm/pull/22168" target="_blank" rel="noopener noreferrer">PR #22168</a></li>
<li>Add logging, spend tracking support + tool tracing - <a href="https://github.com/BerriAI/litellm/pull/22105" target="_blank" rel="noopener noreferrer">PR #22105</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/video_generation">Video Generation</a></strong></p>
<ul>
<li>Add <code>variant</code> parameter to video content download - <a href="https://github.com/BerriAI/litellm/pull/21955" target="_blank" rel="noopener noreferrer">PR #21955</a></li>
<li>Pass <code>api_key</code> from <code>litellm_params</code> to video remix handlers - <a href="https://github.com/BerriAI/litellm/pull/21965" target="_blank" rel="noopener noreferrer">PR #21965</a></li>
<li>Apply custom video pricing from deployment <code>model_info</code> - <a href="https://github.com/BerriAI/litellm/pull/21923" target="_blank" rel="noopener noreferrer">PR #21923</a></li>
<li>Fix passing of image and parameters in videos API - <a href="https://github.com/BerriAI/litellm/pull/22170" target="_blank" rel="noopener noreferrer">PR #22170</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openai#ocr--document-understanding">OCR</a></strong></p>
<ul>
<li>Enable local file support for OCR - <a href="https://github.com/BerriAI/litellm/pull/22133" target="_blank" rel="noopener noreferrer">PR #22133</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/completion/input">Websearch / Tool Calling</a></strong></p>
<ul>
<li>Preserve thinking blocks in agentic loop follow-up messages - <a href="https://github.com/BerriAI/litellm/pull/21604" target="_blank" rel="noopener noreferrer">PR #21604</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Add configurable upper bound for chunk processing time - <a href="https://github.com/BerriAI/litellm/pull/22209" target="_blank" rel="noopener noreferrer">PR #22209</a></li>
<li>Emit <code>x-litellm-overhead-duration-ms</code> header for streaming requests - <a href="https://github.com/BerriAI/litellm/pull/22027" target="_blank" rel="noopener noreferrer">PR #22027</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs">Bugs<a href="https://docs.litellm.ai/release_notes/v1-82-0#bugs" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>General</strong>
<ul>
<li>Fix mypy attr-defined errors on realtime websocket calls - <a href="https://github.com/BerriAI/litellm/pull/22202" target="_blank" rel="noopener noreferrer">PR #22202</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-82-0#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-82-0#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong>Projects</strong></p>
<ul>
<li>Add Projects page with list and create flows - <a href="https://github.com/BerriAI/litellm/pull/22315" target="_blank" rel="noopener noreferrer">PR #22315</a></li>
<li>Add Project Details page with edit modal - <a href="https://github.com/BerriAI/litellm/pull/22360" target="_blank" rel="noopener noreferrer">PR #22360</a></li>
<li>Add project keys table and project dropdown on key create/edit - <a href="https://github.com/BerriAI/litellm/pull/22373" target="_blank" rel="noopener noreferrer">PR #22373</a></li>
<li>Add delete project action to Projects table - <a href="https://github.com/BerriAI/litellm/pull/22412" target="_blank" rel="noopener noreferrer">PR #22412</a></li>
<li>Add Projects Opt-In Toggle in Admin Settings - <a href="https://github.com/BerriAI/litellm/pull/22416" target="_blank" rel="noopener noreferrer">PR #22416</a></li>
<li>Include <code>created_at</code> and <code>updated_at</code> in <code>/project/list</code> response - <a href="https://github.com/BerriAI/litellm/pull/22323" target="_blank" rel="noopener noreferrer">PR #22323</a></li>
<li>Add tags in project - <a href="https://github.com/BerriAI/litellm/pull/22216" target="_blank" rel="noopener noreferrer">PR #22216</a></li>
</ul>
</li>
<li>
<p><strong>Virtual Keys + Access Groups</strong></p>
<ul>
<li>Add bidirectional team/key sync for Access Group CRUD flows - <a href="https://github.com/BerriAI/litellm/pull/22253" target="_blank" rel="noopener noreferrer">PR #22253</a></li>
<li>Add pagination and search to <code>/key/aliases</code> to prevent OOMs - <a href="https://github.com/BerriAI/litellm/pull/22137" target="_blank" rel="noopener noreferrer">PR #22137</a></li>
<li>Add paginated key alias selector in UI - <a href="https://github.com/BerriAI/litellm/pull/22157" target="_blank" rel="noopener noreferrer">PR #22157</a></li>
<li>Add <code>project_id</code> and <code>access_group_id</code> filters for key list endpoint - <a href="https://github.com/BerriAI/litellm/pull/22356" target="_blank" rel="noopener noreferrer">PR #22356</a></li>
<li>Add KeyInfoHeader component - <a href="https://github.com/BerriAI/litellm/pull/22047" target="_blank" rel="noopener noreferrer">PR #22047</a></li>
<li>Restrict Edit Settings to key owners - <a href="https://github.com/BerriAI/litellm/pull/21985" target="_blank" rel="noopener noreferrer">PR #21985</a></li>
<li>Fix virtual key grace period from env/UI - <a href="https://github.com/BerriAI/litellm/pull/20321" target="_blank" rel="noopener noreferrer">PR #20321</a></li>
</ul>
</li>
<li>
<p><strong>Agents</strong></p>
<ul>
<li>Assign virtual keys to agents - <a href="https://github.com/BerriAI/litellm/pull/22045" target="_blank" rel="noopener noreferrer">PR #22045</a></li>
<li>Assign tools to agents - <a href="https://github.com/BerriAI/litellm/pull/22064" target="_blank" rel="noopener noreferrer">PR #22064</a></li>
<li>Ensure internal users cannot create agents (RBAC enforcement) - <a href="https://github.com/BerriAI/litellm/pull/22329" target="_blank" rel="noopener noreferrer">PR #22329</a></li>
</ul>
</li>
<li>
<p><strong>Proxy Auth / SSO</strong></p>
<ul>
<li>OIDC discovery URLs, roles array handling, and dot-notation error hints - <a href="https://github.com/BerriAI/litellm/pull/22336" target="_blank" rel="noopener noreferrer">PR #22336</a></li>
<li>Add PROXY_ADMIN role to system user for key rotation - <a href="https://github.com/BerriAI/litellm/pull/21896" target="_blank" rel="noopener noreferrer">PR #21896</a></li>
</ul>
</li>
<li>
<p><strong>Usage / Spend Logs</strong></p>
<ul>
<li>Add user filtering to usage page - <a href="https://github.com/BerriAI/litellm/pull/22059" target="_blank" rel="noopener noreferrer">PR #22059</a></li>
<li>Allow using AI to understand usage patterns - <a href="https://github.com/BerriAI/litellm/pull/22042" target="_blank" rel="noopener noreferrer">PR #22042</a></li>
<li>Use backend <code>request_duration_ms</code> and make Duration sortable in Logs - <a href="https://github.com/BerriAI/litellm/pull/22122" target="_blank" rel="noopener noreferrer">PR #22122</a></li>
<li>Add <code>request_duration_ms</code> to SpendLogs - <a href="https://github.com/BerriAI/litellm/pull/22066" target="_blank" rel="noopener noreferrer">PR #22066</a></li>
<li>Enrich failure spend logs with key/team metadata - <a href="https://github.com/BerriAI/litellm/pull/22049" target="_blank" rel="noopener noreferrer">PR #22049</a></li>
<li>Show real tool names in logs for Anthropic-format tools - <a href="https://github.com/BerriAI/litellm/pull/22048" target="_blank" rel="noopener noreferrer">PR #22048</a></li>
</ul>
</li>
<li>
<p><strong>Models + Endpoints</strong></p>
<ul>
<li>Show proxy URL in ModelHub - <a href="https://github.com/BerriAI/litellm/pull/21660" target="_blank" rel="noopener noreferrer">PR #21660</a></li>
<li>Add <code>/public/endpoints</code> for provider endpoint support - <a href="https://github.com/BerriAI/litellm/pull/22248" target="_blank" rel="noopener noreferrer">PR #22248</a></li>
</ul>
</li>
<li>
<p><strong>UI Improvements</strong></p>
<ul>
<li>Add custom favicon support - <a href="https://github.com/BerriAI/litellm/pull/21653" target="_blank" rel="noopener noreferrer">PR #21653</a></li>
<li>Add Blog Dropdown in Navbar - <a href="https://github.com/BerriAI/litellm/pull/21859" target="_blank" rel="noopener noreferrer">PR #21859</a></li>
<li>Add UI banner warning for detailed debug mode - <a href="https://github.com/BerriAI/litellm/pull/21527" target="_blank" rel="noopener noreferrer">PR #21527</a></li>
<li>Make auth value optional for MCP Server create flow - <a href="https://github.com/BerriAI/litellm/pull/22119" target="_blank" rel="noopener noreferrer">PR #22119</a></li>
<li>Tool policies: auto-discover tools + policy enforcement guardrail - <a href="https://github.com/BerriAI/litellm/pull/22041" target="_blank" rel="noopener noreferrer">PR #22041</a></li>
</ul>
</li>
<li>
<p><strong>Health Checks</strong></p>
<ul>
<li>Add health check max tokens configuration - <a href="https://github.com/BerriAI/litellm/pull/22299" target="_blank" rel="noopener noreferrer">PR #22299</a></li>
<li>Limit concurrent health checks with <code>health_check_concurrency</code> - <a href="https://github.com/BerriAI/litellm/pull/20584" target="_blank" rel="noopener noreferrer">PR #20584</a></li>
<li>Fix health check <code>model_id</code> filtering - <a href="https://github.com/BerriAI/litellm/pull/21071" target="_blank" rel="noopener noreferrer">PR #21071</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs-1">Bugs<a href="https://docs.litellm.ai/release_notes/v1-82-0#bugs-1" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li>Populate <code>user_id</code> and <code>user_info</code> for admin users in <code>/user/info</code> - <a href="https://github.com/BerriAI/litellm/pull/22239" target="_blank" rel="noopener noreferrer">PR #22239</a></li>
<li>Fix virtual keys pagination stale totals when filtering - <a href="https://github.com/BerriAI/litellm/pull/22222" target="_blank" rel="noopener noreferrer">PR #22222</a></li>
<li>Fix Spend Update Queue aggregation never triggers with default presets - <a href="https://github.com/BerriAI/litellm/pull/21963" target="_blank" rel="noopener noreferrer">PR #21963</a></li>
<li>Fix timezone config lookup and replace hardcoded timezone map with <code>ZoneInfo</code> - <a href="https://github.com/BerriAI/litellm/pull/21754" target="_blank" rel="noopener noreferrer">PR #21754</a></li>
<li>Fix custom auth budget issue - <a href="https://github.com/BerriAI/litellm/pull/22164" target="_blank" rel="noopener noreferrer">PR #22164</a></li>
<li>Fix missing OAuth session state - <a href="https://github.com/BerriAI/litellm/pull/21992" target="_blank" rel="noopener noreferrer">PR #21992</a></li>
<li>Fix Transport Type for OpenAPI Spec on UI - <a href="https://github.com/BerriAI/litellm/pull/22005" target="_blank" rel="noopener noreferrer">PR #22005</a></li>
<li>Fix Claude Code plugin schema - <a href="https://github.com/BerriAI/litellm/pull/22271" target="_blank" rel="noopener noreferrer">PR #22271</a></li>
<li>Add missing migration for <code>LiteLLM_ClaudeCodePluginTable</code> - <a href="https://github.com/BerriAI/litellm/pull/22335" target="_blank" rel="noopener noreferrer">PR #22335</a></li>
<li>Only tag selected deployment in access group creation - <a href="https://github.com/BerriAI/litellm/pull/21655" target="_blank" rel="noopener noreferrer">PR #21655</a></li>
<li>State management fixes for CheckBatchCost - <a href="https://github.com/BerriAI/litellm/pull/21921" target="_blank" rel="noopener noreferrer">PR #21921</a></li>
<li>Remove duplicate antd import in ToolPolicies - <a href="https://github.com/BerriAI/litellm/pull/22107" target="_blank" rel="noopener noreferrer">PR #22107</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="ai-integrations">AI Integrations<a href="https://docs.litellm.ai/release_notes/v1-82-0#ai-integrations" class="hash-link" aria-label="Direct link to AI Integrations" title="Direct link to AI Integrations">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="logging">Logging<a href="https://docs.litellm.ai/release_notes/v1-82-0#logging" class="hash-link" aria-label="Direct link to Logging" title="Direct link to Logging">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#datadog">DataDog</a></strong></p>
<ul>
<li>Add ability to trace metrics in DataDog - <a href="https://github.com/BerriAI/litellm/pull/22103" target="_blank" rel="noopener noreferrer">PR #22103</a></li>
<li>Correlate LiteLLM call IDs with DataDog APM spans - <a href="https://github.com/BerriAI/litellm/pull/22219" target="_blank" rel="noopener noreferrer">PR #22219</a></li>
<li>Fix TTS metric emission issues - <a href="https://github.com/BerriAI/litellm/pull/20632" target="_blank" rel="noopener noreferrer">PR #20632</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#prometheus">Prometheus</a></strong></p>
<ul>
<li>Add opt-in <code>stream</code> label on <code>litellm_proxy_total_requests_metric</code> - <a href="https://github.com/BerriAI/litellm/pull/22023" target="_blank" rel="noopener noreferrer">PR #22023</a></li>
<li>Fix team <code>+Inf</code> budgets in Prometheus metrics - <a href="https://github.com/BerriAI/litellm/pull/22243" target="_blank" rel="noopener noreferrer">PR #22243</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#langfuse">Langfuse</a></strong></p>
<ul>
<li>Fix Langfuse OTEL trace issues - <a href="https://github.com/BerriAI/litellm/pull/21309" target="_blank" rel="noopener noreferrer">PR #21309</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/observability/arize_phoenix">Arize Phoenix</a></strong></p>
<ul>
<li>Fix nested traces coexistence with OTEL callback - <a href="https://github.com/BerriAI/litellm/pull/22169" target="_blank" rel="noopener noreferrer">PR #22169</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/alerting">Slack</a></strong></p>
<ul>
<li>Add optional digest mode for Slack alert types - <a href="https://github.com/BerriAI/litellm/pull/21683" target="_blank" rel="noopener noreferrer">PR #21683</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Fix Gemini trace ID missing in logging - <a href="https://github.com/BerriAI/litellm/pull/22077" target="_blank" rel="noopener noreferrer">PR #22077</a></li>
<li>Populate <code>cache_read_input_tokens</code> from <code>prompt_tokens_details</code> for OpenAI/Azure - <a href="https://github.com/BerriAI/litellm/pull/22090" target="_blank" rel="noopener noreferrer">PR #22090</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails">Guardrails<a href="https://docs.litellm.ai/release_notes/v1-82-0#guardrails" class="hash-link" aria-label="Direct link to Guardrails" title="Direct link to Guardrails">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">Noma</a></strong></p>
<ul>
<li>Noma guardrails v2 based on custom guardrails framework - <a href="https://github.com/BerriAI/litellm/pull/21400" target="_blank" rel="noopener noreferrer">PR #21400</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">LakeraAI</a></strong></p>
<ul>
<li>Add Lakera v2 post-call hook with fixed PII masking - <a href="https://github.com/BerriAI/litellm/pull/21783" target="_blank" rel="noopener noreferrer">PR #21783</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">Presidio</a></strong></p>
<ul>
<li>Fix Presidio streaming and false positives - <a href="https://github.com/BerriAI/litellm/pull/21949" target="_blank" rel="noopener noreferrer">PR #21949</a></li>
<li>Fix Presidio streaming v3 reliability improvements - <a href="https://github.com/BerriAI/litellm/pull/22283" target="_blank" rel="noopener noreferrer">PR #22283</a></li>
<li>Prevent Presidio crash on non-JSON responses - <a href="https://github.com/BerriAI/litellm/pull/22084" target="_blank" rel="noopener noreferrer">PR #22084</a></li>
</ul>
</li>
<li>
<p><strong>Built-in Guardrails</strong></p>
<ul>
<li>Block code execution guardrail to prevent agents from executing code - <a href="https://github.com/BerriAI/litellm/pull/22154" target="_blank" rel="noopener noreferrer">PR #22154</a></li>
<li>Employment discrimination topic blockers for 5 protected classes - <a href="https://github.com/BerriAI/litellm/pull/21962" target="_blank" rel="noopener noreferrer">PR #21962</a></li>
<li>Claims agent guardrails (5 categories + policy template) - <a href="https://github.com/BerriAI/litellm/pull/22113" target="_blank" rel="noopener noreferrer">PR #22113</a></li>
<li>New code execution evaluation dataset - <a href="https://github.com/BerriAI/litellm/pull/22065" target="_blank" rel="noopener noreferrer">PR #22065</a></li>
<li>Tool policies: auto-discover tools + policy enforcement - <a href="https://github.com/BerriAI/litellm/pull/22041" target="_blank" rel="noopener noreferrer">PR #22041</a></li>
</ul>
</li>
<li>
<p><strong>Policy Templates</strong></p>
<ul>
<li>Singapore guardrail policies (PDPA + MAS AI Risk Management) - <a href="https://github.com/BerriAI/litellm/pull/21948" target="_blank" rel="noopener noreferrer">PR #21948</a></li>
<li>Prefix SG guardrail policy IDs with country code - <a href="https://github.com/BerriAI/litellm/pull/21974" target="_blank" rel="noopener noreferrer">PR #21974</a></li>
<li>Guardrail policy versioning - <a href="https://github.com/BerriAI/litellm/pull/21862" target="_blank" rel="noopener noreferrer">PR #21862</a></li>
</ul>
</li>
<li>
<p><strong>Guardrail Monitoring</strong></p>
<ul>
<li>Guardrail Monitor — measure guardrail reliability in production - <a href="https://github.com/BerriAI/litellm/pull/21944" target="_blank" rel="noopener noreferrer">PR #21944</a></li>
</ul>
</li>
<li>
<p><strong>Security</strong></p>
<ul>
<li>Fix unauthenticated RCE and sandbox escape in custom code guardrail - <a href="https://github.com/BerriAI/litellm/pull/22095" target="_blank" rel="noopener noreferrer">PR #22095</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="prompt-management">Prompt Management<a href="https://docs.litellm.ai/release_notes/v1-82-0#prompt-management" class="hash-link" aria-label="Direct link to Prompt Management" title="Direct link to Prompt Management">​</a></h3>
<p>No major prompt management changes in this release.</p>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="secret-managers">Secret Managers<a href="https://docs.litellm.ai/release_notes/v1-82-0#secret-managers" class="hash-link" aria-label="Direct link to Secret Managers" title="Direct link to Secret Managers">​</a></h3>
<p>No major secret manager changes in this release.</p>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-82-0#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li><strong>Priority PayGo cost tracking</strong> for Gemini/Vertex AI - <a href="https://github.com/BerriAI/litellm/pull/21909" target="_blank" rel="noopener noreferrer">PR #21909</a></li>
<li><strong>Add <code>request_duration_ms</code> to SpendLogs</strong> for latency tracking per request - <a href="https://github.com/BerriAI/litellm/pull/22066" target="_blank" rel="noopener noreferrer">PR #22066</a></li>
<li><strong>Add <code>in_flight_requests</code> metric</strong> to <code>/health/backlog</code> + Prometheus - <a href="https://github.com/BerriAI/litellm/pull/22319" target="_blank" rel="noopener noreferrer">PR #22319</a></li>
<li><strong>Enrich failure spend logs</strong> with key/team metadata - <a href="https://github.com/BerriAI/litellm/pull/22049" target="_blank" rel="noopener noreferrer">PR #22049</a></li>
<li><strong>Add spend tracking lifecycle logging</strong> for debugging spend flows - <a href="https://github.com/BerriAI/litellm/pull/22029" target="_blank" rel="noopener noreferrer">PR #22029</a></li>
<li><strong>Fix budget timezone config lookup</strong> and replace hardcoded timezone map with <code>ZoneInfo</code> - <a href="https://github.com/BerriAI/litellm/pull/21754" target="_blank" rel="noopener noreferrer">PR #21754</a></li>
<li><strong>Fix Spend Update Queue aggregation</strong> never triggering with default presets - <a href="https://github.com/BerriAI/litellm/pull/21963" target="_blank" rel="noopener noreferrer">PR #21963</a></li>
<li><strong>Avoid mutating caller-owned dicts</strong> in <code>SpendUpdateQueue</code> aggregation - <a href="https://github.com/BerriAI/litellm/pull/21742" target="_blank" rel="noopener noreferrer">PR #21742</a></li>
<li><strong>Optimize old spendlog deletion</strong> cron job - <a href="https://github.com/BerriAI/litellm/pull/21930" target="_blank" rel="noopener noreferrer">PR #21930</a></li>
<li><strong>Health check max tokens</strong> configuration - <a href="https://github.com/BerriAI/litellm/pull/22299" target="_blank" rel="noopener noreferrer">PR #22299</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-gateway">MCP Gateway<a href="https://docs.litellm.ai/release_notes/v1-82-0#mcp-gateway" class="hash-link" aria-label="Direct link to MCP Gateway" title="Direct link to MCP Gateway">​</a></h2>
<ul>
<li><strong>Pass MCP auth headers</strong> from request context to tool fetch for <code>/v1/responses</code> and <code>/chat/completions</code> - <a href="https://github.com/BerriAI/litellm/pull/22291" target="_blank" rel="noopener noreferrer">PR #22291</a></li>
<li><strong>Default <code>available_on_public_internet</code> to true</strong> for MCP server behavior consistency - <a href="https://github.com/BerriAI/litellm/pull/22331" target="_blank" rel="noopener noreferrer">PR #22331</a></li>
<li><strong>Clear error messages</strong> for IP filtering / no available tools - <a href="https://github.com/BerriAI/litellm/pull/22142" target="_blank" rel="noopener noreferrer">PR #22142</a></li>
<li><strong>Strip stale <code>mcp-session-id</code> header</strong> to prevent 400 errors across proxy workers - <a href="https://github.com/BerriAI/litellm/pull/21417" target="_blank" rel="noopener noreferrer">PR #21417</a></li>
<li><strong>Skip health check for MCP</strong> with passthrough token auth - <a href="https://github.com/BerriAI/litellm/pull/21982" target="_blank" rel="noopener noreferrer">PR #21982</a></li>
<li><strong>Fix missing OAuth session state</strong> - <a href="https://github.com/BerriAI/litellm/pull/21992" target="_blank" rel="noopener noreferrer">PR #21992</a></li>
<li><strong>Fix Transport Type</strong> for OpenAPI Spec on UI - <a href="https://github.com/BerriAI/litellm/pull/22005" target="_blank" rel="noopener noreferrer">PR #22005</a></li>
<li><strong>Add e2e test</strong> for stateless StreamableHTTP behavior - <a href="https://github.com/BerriAI/litellm/pull/22033" target="_blank" rel="noopener noreferrer">PR #22033</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements">Performance / Loadbalancing / Reliability improvements<a href="https://docs.litellm.ai/release_notes/v1-82-0#performance--loadbalancing--reliability-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements" title="Direct link to Performance / Loadbalancing / Reliability improvements">​</a></h2>
<p><strong>Streaming &amp; hot-path</strong></p>
<ul>
<li>Streaming latency improvements — 4 targeted hot-path fixes - <a href="https://github.com/BerriAI/litellm/pull/22346" target="_blank" rel="noopener noreferrer">PR #22346</a></li>
<li>Skip throwaway <code>Usage()</code> construction in <code>ModelResponse.__init__</code> - <a href="https://github.com/BerriAI/litellm/pull/21611" target="_blank" rel="noopener noreferrer">PR #21611</a></li>
<li>Optimize <code>is_model_o_series_model</code> with <code>startswith</code> - <a href="https://github.com/BerriAI/litellm/pull/21690" target="_blank" rel="noopener noreferrer">PR #21690</a></li>
<li>Use cached <code>_safe_get_request_headers</code> instead of per-request construction - <a href="https://github.com/BerriAI/litellm/pull/21430" target="_blank" rel="noopener noreferrer">PR #21430</a></li>
<li>Emit <code>x-litellm-overhead-duration-ms</code> header for streaming requests - <a href="https://github.com/BerriAI/litellm/pull/22027" target="_blank" rel="noopener noreferrer">PR #22027</a></li>
</ul>
<p><strong>Database &amp; Redis</strong></p>
<ul>
<li>Batch 11 <code>create_task()</code> calls into 1 in <code>update_database()</code> - <a href="https://github.com/BerriAI/litellm/pull/22028" target="_blank" rel="noopener noreferrer">PR #22028</a></li>
<li>Redis pipeline spend updates for batched writes - <a href="https://github.com/BerriAI/litellm/pull/22044" target="_blank" rel="noopener noreferrer">PR #22044</a></li>
<li>Recover from prisma-query-engine zombie process - <a href="https://github.com/BerriAI/litellm/pull/21899" target="_blank" rel="noopener noreferrer">PR #21899</a></li>
<li>Optimize old spendlog deletion cron job - <a href="https://github.com/BerriAI/litellm/pull/21930" target="_blank" rel="noopener noreferrer">PR #21930</a></li>
</ul>
<p><strong>Router &amp; caching</strong></p>
<ul>
<li>Add cache invalidation for <code>_cached_get_model_group_info</code> - <a href="https://github.com/BerriAI/litellm/pull/20376" target="_blank" rel="noopener noreferrer">PR #20376</a></li>
<li>Remove cache eviction close that kills in-use httpx clients - <a href="https://github.com/BerriAI/litellm/pull/22247" target="_blank" rel="noopener noreferrer">PR #22247</a></li>
<li>Store background task references in <code>LLMClientCache._remove_key</code> to prevent unawaited coroutine warnings - <a href="https://github.com/BerriAI/litellm/pull/22143" target="_blank" rel="noopener noreferrer">PR #22143</a></li>
<li>Fix <code>ensure_arrival_time</code> set before calculating queue time - <a href="https://github.com/BerriAI/litellm/pull/21918" target="_blank" rel="noopener noreferrer">PR #21918</a></li>
</ul>
<p><strong>Connection management</strong></p>
<ul>
<li>Only set <code>enable_cleanup_closed</code> on aiohttp when required - <a href="https://github.com/BerriAI/litellm/pull/21897" target="_blank" rel="noopener noreferrer">PR #21897</a></li>
<li>Prometheus child_exit cleanup for gunicorn workers - <a href="https://github.com/BerriAI/litellm/pull/22324" target="_blank" rel="noopener noreferrer">PR #22324</a></li>
<li>Prometheus multiprocess cleanup - <a href="https://github.com/BerriAI/litellm/pull/22221" target="_blank" rel="noopener noreferrer">PR #22221</a></li>
<li>Limit concurrent health checks with <code>health_check_concurrency</code> - <a href="https://github.com/BerriAI/litellm/pull/20584" target="_blank" rel="noopener noreferrer">PR #20584</a></li>
<li>Isolate <code>get_config</code> failures from model sync loop - <a href="https://github.com/BerriAI/litellm/pull/22224" target="_blank" rel="noopener noreferrer">PR #22224</a></li>
</ul>
<p><strong>Other</strong></p>
<ul>
<li>Semantic cache: support configurable vector dimensions - <a href="https://github.com/BerriAI/litellm/pull/21649" target="_blank" rel="noopener noreferrer">PR #21649</a></li>
<li>Honor <code>MAX_STRING_LENGTH_PROMPT_IN_DB</code> from config env vars - <a href="https://github.com/BerriAI/litellm/pull/22106" target="_blank" rel="noopener noreferrer">PR #22106</a></li>
<li>Enhance <code>MidStreamFallbackError</code> to preserve original status code and attributes - <a href="https://github.com/BerriAI/litellm/pull/22225" target="_blank" rel="noopener noreferrer">PR #22225</a></li>
<li>Network mock utility for testing - <a href="https://github.com/BerriAI/litellm/pull/21942" target="_blank" rel="noopener noreferrer">PR #21942</a></li>
<li>Add missing return type annotations to iterator protocol methods in streaming_handler - <a href="https://github.com/BerriAI/litellm/pull/21750" target="_blank" rel="noopener noreferrer">PR #21750</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="security">Security<a href="https://docs.litellm.ai/release_notes/v1-82-0#security" class="hash-link" aria-label="Direct link to Security" title="Direct link to Security">​</a></h2>
<ul>
<li>Fix critical/high CVEs in OS-level libs and NPM transitive dependencies - <a href="https://github.com/BerriAI/litellm/pull/22008" target="_blank" rel="noopener noreferrer">PR #22008</a></li>
<li>Fix unauthenticated RCE and sandbox escape in custom code guardrail - <a href="https://github.com/BerriAI/litellm/pull/22095" target="_blank" rel="noopener noreferrer">PR #22095</a></li>
<li>Remove hardcoded base64 string flagged by secret scanner - <a href="https://github.com/BerriAI/litellm/pull/22125" target="_blank" rel="noopener noreferrer">PR #22125</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates">Documentation Updates<a href="https://docs.litellm.ai/release_notes/v1-82-0#documentation-updates" class="hash-link" aria-label="Direct link to Documentation Updates" title="Direct link to Documentation Updates">​</a></h2>
<ul>
<li>Add OpenAI Agents SDK tutorial with LiteLLM Proxy - <a href="https://github.com/BerriAI/litellm/pull/21221" target="_blank" rel="noopener noreferrer">PR #21221</a></li>
<li>Add OpenClaw integration tutorial - <a href="https://github.com/BerriAI/litellm/pull/21605" target="_blank" rel="noopener noreferrer">PR #21605</a></li>
<li>Add Google GenAI SDK tutorial (JS &amp; Python) - <a href="https://github.com/BerriAI/litellm/pull/21885" target="_blank" rel="noopener noreferrer">PR #21885</a></li>
<li>Add Gollem Go agent framework cookbook example - <a href="https://github.com/BerriAI/litellm/pull/21747" target="_blank" rel="noopener noreferrer">PR #21747</a></li>
<li>Update AssemblyAI docs with Universal-3 Pro, Speech Understanding, and LLM Gateway - <a href="https://github.com/BerriAI/litellm/pull/21130" target="_blank" rel="noopener noreferrer">PR #21130</a></li>
<li>Add <code>store_model_in_db</code> release docs - <a href="https://github.com/BerriAI/litellm/pull/21863" target="_blank" rel="noopener noreferrer">PR #21863</a></li>
<li>Add Credential Usage Tracking docs - <a href="https://github.com/BerriAI/litellm/pull/22112" target="_blank" rel="noopener noreferrer">PR #22112</a></li>
<li>Add proxy request tags docs - <a href="https://github.com/BerriAI/litellm/pull/22129" target="_blank" rel="noopener noreferrer">PR #22129</a></li>
<li>Add trailing slash to <code>/mcp</code> endpoint URLs - <a href="https://github.com/BerriAI/litellm/pull/20509" target="_blank" rel="noopener noreferrer">PR #20509</a></li>
<li>Add pre-PR checklist to UI contributing guide - <a href="https://github.com/BerriAI/litellm/pull/21886" target="_blank" rel="noopener noreferrer">PR #21886</a></li>
<li>Replace Azure OpenAI key with mock key in docs - <a href="https://github.com/BerriAI/litellm/pull/21997" target="_blank" rel="noopener noreferrer">PR #21997</a></li>
<li>Add performance &amp; reliability section to v1.81.14 release notes - <a href="https://github.com/BerriAI/litellm/pull/21950" target="_blank" rel="noopener noreferrer">PR #21950</a></li>
<li>Update v1.81.12-stable release notes to point to stable.1 - <a href="https://github.com/BerriAI/litellm/pull/22036" target="_blank" rel="noopener noreferrer">PR #22036</a></li>
<li>Add security vulnerability scan report to v1.81.14 release notes - <a href="https://github.com/BerriAI/litellm/pull/22385" target="_blank" rel="noopener noreferrer">PR #22385</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-82-0#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@janfrederickk made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21660" target="_blank" rel="noopener noreferrer">PR #21660</a></li>
<li>@hztBUAA made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21656" target="_blank" rel="noopener noreferrer">PR #21656</a></li>
<li>@LeeJuOh made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21754" target="_blank" rel="noopener noreferrer">PR #21754</a></li>
<li>@WhoisMonesh made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21750" target="_blank" rel="noopener noreferrer">PR #21750</a></li>
<li>@trevorprater made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21747" target="_blank" rel="noopener noreferrer">PR #21747</a></li>
<li>@edwiniac made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21870" target="_blank" rel="noopener noreferrer">PR #21870</a></li>
<li>@stakeswky made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21867" target="_blank" rel="noopener noreferrer">PR #21867</a></li>
<li>@ta-stripe made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21701" target="_blank" rel="noopener noreferrer">PR #21701</a></li>
<li>@ron-zhong made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21948" target="_blank" rel="noopener noreferrer">PR #21948</a></li>
<li>@Arindam200 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21221" target="_blank" rel="noopener noreferrer">PR #21221</a></li>
<li>@Canvinus made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21964" target="_blank" rel="noopener noreferrer">PR #21964</a></li>
<li>@nicolopignatelli made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21951" target="_blank" rel="noopener noreferrer">PR #21951</a></li>
<li>@MarshHawk made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20584" target="_blank" rel="noopener noreferrer">PR #20584</a></li>
<li>@gavksingh made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/22106" target="_blank" rel="noopener noreferrer">PR #22106</a></li>
<li>@roni-frantchi made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/22090" target="_blank" rel="noopener noreferrer">PR #22090</a></li>
<li>@noahnistler made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/22133" target="_blank" rel="noopener noreferrer">PR #22133</a></li>
<li>@dylan-duan-aai made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21130" target="_blank" rel="noopener noreferrer">PR #21130</a></li>
<li>@rasmi made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/22322" target="_blank" rel="noopener noreferrer">PR #22322</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="diff-summary">Diff Summary<a href="https://docs.litellm.ai/release_notes/v1-82-0#diff-summary" class="hash-link" aria-label="Direct link to Diff Summary" title="Direct link to Diff Summary">​</a></h2>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="02282026">02/28/2026<a href="https://docs.litellm.ai/release_notes/v1-82-0#02282026" class="hash-link" aria-label="Direct link to 02/28/2026" title="Direct link to 02/28/2026">​</a></h2>
<ul>
<li>New Models / Updated Models: 26</li>
<li>LLM API Endpoints: 14</li>
<li>Management Endpoints / UI: 38</li>
<li>AI Integrations: 25</li>
<li>Spend Tracking, Budgets and Rate Limiting: 10</li>
<li>MCP Gateway: 8</li>
<li>Performance / Loadbalancing / Reliability improvements: 22</li>
<li>Security: 3</li>
<li>Documentation Updates: 14</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog">Full Changelog<a href="https://docs.litellm.ai/release_notes/v1-82-0#full-changelog" class="hash-link" aria-label="Direct link to Full Changelog" title="Direct link to Full Changelog">​</a></h2>
<p><a href="https://github.com/BerriAI/litellm/compare/v1.81.14.rc.1...v1.82.0" target="_blank" rel="noopener noreferrer">v1.81.14.rc.1...v1.82.0</a></p>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[v1.81.14 - New Gateway Level Guardrails & Compliance Playground]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-81-14</id>
        <link href="https://docs.litellm.ai/release_notes/v1-81-14"/>
        <updated>2026-02-21T00:00:00.000Z</updated>
        <summary type="html"><![CDATA[Deploy this version]]></summary>
        <content type="html"><![CDATA[<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-81-14#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<!-- -->
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">ghcr.io/berriai/litellm:main-v1.81.14-stable</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.81.14</span><br></span></code></pre></div></div></div></div></div>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-81-14#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<ul>
<li><strong>Guardrail Garden</strong> — <a href="https://docs.litellm.ai/docs/proxy/guardrails/policy_templates">Browse built-in and partner guardrails by use case — competitor blocking, topic filtering, GDPR, prompt injection, and more. Pick a template, customize it, attach it to a team or key.</a></li>
<li><strong>Compliance Playground</strong> — <a href="https://docs.litellm.ai/docs/proxy/guardrails/policy_templates">Test any guardrail policy against your own traffic before it goes live. See precision, recall, and false positive rate — so you know how it'll behave in production.</a></li>
<li><strong>3 new zero-cost built-in guardrails</strong> — <a href="https://docs.litellm.ai/docs/proxy/guardrails">Competitor name blocker, topic blocker, and insults filter — all gateway-level, &lt;0.1ms latency, no external API, configurable per-team or key</a></li>
<li><strong>Store Model in DB Settings via UI</strong> - <a href="https://docs.litellm.ai/docs/proxy/ui_store_model_db_setting">Configure model storage directly in the Admin UI without editing config files or restarting the proxy—perfect for cloud deployments</a></li>
<li><strong>Claude Sonnet 4.6 — day 0</strong> — <a href="https://docs.litellm.ai/docs/providers/anthropic">Full support across Anthropic and Vertex AI: reasoning, computer use, prompt caching, 200K context</a></li>
<li><strong>20+ performance optimizations</strong> — Faster routing, lower logging overhead, reduced cost-calculator latency, and connection pool fixes — meaningfully less CPU and latency on every request</li>
</ul>
<hr>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="guardrail-garden">Guardrail Garden<a href="https://docs.litellm.ai/release_notes/v1-81-14#guardrail-garden" class="hash-link" aria-label="Direct link to Guardrail Garden" title="Direct link to Guardrail Garden">​</a></h3>
<p>AI Platform Admins can now browse built-in and partner guardrails from the Guardrail Garden. Guardrails are organized by use case — blocking financial advice, filtering insults, detecting competitor mentions, and more — so you can find the right one and deploy it in a few clicks.</p>
<p><img decoding="async" loading="lazy" alt="Guardrail Garden" src="https://docs.litellm.ai/assets/images/guardrail_garden-17e0f53e4413c8ef4409721b0a177df0.png" width="4260" height="2508" class="img_ev3q"></p>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="3-new-built-in-guardrails">3 New Built-in Guardrails<a href="https://docs.litellm.ai/release_notes/v1-81-14#3-new-built-in-guardrails" class="hash-link" aria-label="Direct link to 3 New Built-in Guardrails" title="Direct link to 3 New Built-in Guardrails">​</a></h3>
<p>This release brings 3 new built-in guardrails that run directly on the gateway. This is great for AI Gateway Admins who need low latency, zero cost guardrails for their scenarios.</p>
<ul>
<li><strong>Denied Financial Advice</strong> — detects requests for personalized financial advice, investment recommendations, or financial planning</li>
<li><strong>Denied Insults</strong> — detects insults, name-calling, and personal attacks directed at the chatbot, staff, or other people</li>
<li><strong>Competitor Name Blocker</strong> — detects mentions of competitor brands in responses</li>
</ul>
<p>These guardrails are built for production and on our benchmarks had a 100% Recall and Precision.</p>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="store-model-in-db-settings-via-ui">Store Model in DB Settings via UI<a href="https://docs.litellm.ai/release_notes/v1-81-14#store-model-in-db-settings-via-ui" class="hash-link" aria-label="Direct link to Store Model in DB Settings via UI" title="Direct link to Store Model in DB Settings via UI">​</a></h3>
<p>Previously, the <code>store_model_in_db</code> setting could only be configured in <code>proxy_config.yaml</code> under <code>general_settings</code>, requiring a proxy restart to take effect. Now you can enable or disable this setting directly from the Admin UI without any restarts. This is especially useful for cloud deployments where you don't have direct access to config files or want to avoid downtime. Enable <code>store_model_in_db</code> to move model definitions from your YAML into the database—reducing config complexity, improving scalability, and enabling dynamic model management across multiple proxy instances.</p>
<p><img decoding="async" loading="lazy" alt="Store model in DB Setting" src="https://docs.litellm.ai/assets/images/ui_store_model_in_db-4a33798f081424f47f4eec65f3c4684c.png" width="4800" height="2508" class="img_ev3q"></p>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="eval-results">Eval results<a href="https://docs.litellm.ai/release_notes/v1-81-14#eval-results" class="hash-link" aria-label="Direct link to Eval results" title="Direct link to Eval results">​</a></h4>
<p>We benchmarked our new built-in guardrails against labeled datasets before shipping. You can see the results for Denied Financial Advice (207 cases) and Denied Insults (299 cases):</p>
<table><thead><tr><th>Guardrail</th><th>Precision</th><th>Recall</th><th>F1</th><th>Latency p50</th><th>Cost/req</th></tr></thead><tbody><tr><td>Denied Financial Advice</td><td>100%</td><td>100%</td><td>100%</td><td>&lt;0.1ms</td><td>$0</td></tr><tr><td>Denied Insults</td><td>100%</td><td>100%</td><td>100%</td><td>&lt;0.1ms</td><td>$0</td></tr></tbody></table>
<p>100% precision means zero false positives — no legitimate messages were incorrectly blocked. 100% recall means zero false negatives — every message that should have been blocked was caught.</p>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="compliance-playground">Compliance Playground<a href="https://docs.litellm.ai/release_notes/v1-81-14#compliance-playground" class="hash-link" aria-label="Direct link to Compliance Playground" title="Direct link to Compliance Playground">​</a></h3>
<p>The Compliance Playground lets you test any guardrail against our pre-built eval datasets or your own custom datasets, so you can see precision, recall, and false positive rate before rolling it out to production.</p>
<p><img decoding="async" loading="lazy" alt="Compliance Playground" src="https://docs.litellm.ai/assets/images/compliance_playground-11b8b8246548c5a015746fad79f20cff.png" width="4260" height="2508" class="img_ev3q"></p>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--reliability--up-to-13-lower-latency">Performance &amp; Reliability — Up to 13% Lower Latency<a href="https://docs.litellm.ai/release_notes/v1-81-14#performance--reliability--up-to-13-lower-latency" class="hash-link" aria-label="Direct link to Performance &amp; Reliability — Up to 13% Lower Latency" title="Direct link to Performance &amp; Reliability — Up to 13% Lower Latency">​</a></h2>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAGCAYAAAD68A/GAAAACXBIWXMAABYlAAAWJQFJUiTwAAAAuklEQVR4nD3P3UrDQBDF8X0KTXYmH9qGZJM2Sd22aWuJWBC9UCuC7/8of9kVvPhdDOdwYIy1lqIouStLmrrGNe6Pc6gqIRcRTHqTsDpuOPy88PAx4z9n/PWJqnekSYpkiuYZJhxN39KfPe000B3GaFlXqChWLKKCsaG4XePfZ8bXE5u3x2jhKkL2v5jcJtTrluG8pd0PdNNINw3cLxeoSFzTTDFFXrI6enbfF3bXZ/Zfl6jqwjMZmufRL/RkWKJEQRmMAAAAAElFTkSuQmCC&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="369"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/v1_81_14_perf.f9397ca.640.png" srcset="/assets/ideal-img/v1_81_14_perf.f9397ca.640.png 640w,/assets/ideal-img/v1_81_14_perf.7c3ad1c.1920.png 1920w" width="640" height="369"></noscript></div>
<p>This release cuts latency across all percentiles through 20+ micro-optimizations across logging, cost calculation, routing, and connection management. See <a href="https://docs.litellm.ai/docs/benchmarks">benchmarking</a> for more info about how to benchmark yourself.</p>
<ul>
<li><strong>Mean latency:</strong> 78.4 ms → <strong>70.3 ms</strong> (−10.3%)</li>
<li><strong>p50 latency:</strong> 64.8 ms → <strong>57.3 ms</strong> (−11.7%)</li>
<li><strong>p99 latency:</strong> 288.9 ms → <strong>250.0 ms</strong> (−13.4%)</li>
</ul>
<p><strong>Streaming Connection Pool Fix</strong></p>
<p>Fixed a 3-fold connection leak that caused TCP connection starvation under streaming workloads: the aiohttp transport wasn't closing connections, no <code>finally</code> blocks were calling close on disconnect, and a Uvicorn bug prevented disconnect signaling. <a href="https://github.com/BerriAI/litellm/pull/21213" target="_blank" rel="noopener noreferrer">PR #21213</a></p>
<!-- -->
<p><strong>Redis Connection Pool Reliability</strong></p>
<p>Fixed 4 separate connection pool bugs to make how we use Redis more reliable. The most important change was on pools being leaked on cache expiry and the other fixes are detailed here in <a href="https://github.com/BerriAI/litellm/pull/21717" target="_blank" rel="noopener noreferrer">PR #21717</a>.</p>
<!-- -->
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-providers-and-endpoints">New Providers and Endpoints<a href="https://docs.litellm.ai/release_notes/v1-81-14#new-providers-and-endpoints" class="hash-link" aria-label="Direct link to New Providers and Endpoints" title="Direct link to New Providers and Endpoints">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="new-providers-1-new-provider">New Providers (1 new provider)<a href="https://docs.litellm.ai/release_notes/v1-81-14#new-providers-1-new-provider" class="hash-link" aria-label="Direct link to New Providers (1 new provider)" title="Direct link to New Providers (1 new provider)">​</a></h3>
<table><thead><tr><th>Provider</th><th>Supported LiteLLM Endpoints</th><th>Description</th></tr></thead><tbody><tr><td><a href="https://docs.litellm.ai/docs/providers/watsonx">IBM watsonx.ai</a></td><td><code>/rerank</code></td><td>Rerank support for IBM watsonx.ai models</td></tr></tbody></table>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="new-llm-api-endpoints-1-new-endpoint">New LLM API Endpoints (1 new endpoint)<a href="https://docs.litellm.ai/release_notes/v1-81-14#new-llm-api-endpoints-1-new-endpoint" class="hash-link" aria-label="Direct link to New LLM API Endpoints (1 new endpoint)" title="Direct link to New LLM API Endpoints (1 new endpoint)">​</a></h3>
<table><thead><tr><th>Endpoint</th><th>Method</th><th>Description</th><th>Documentation</th></tr></thead><tbody><tr><td><code>/v1/evals</code></td><td>POST/GET</td><td>OpenAI-compatible Evals API for model evaluation</td><td><a href="https://docs.litellm.ai/docs/evals_api">Docs</a></td></tr></tbody></table>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-81-14#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support-13-new-models">New Model Support (13 new models)<a href="https://docs.litellm.ai/release_notes/v1-81-14#new-model-support-13-new-models" class="hash-link" aria-label="Direct link to New Model Support (13 new models)" title="Direct link to New Model Support (13 new models)">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th><th>Features</th></tr></thead><tbody><tr><td>Anthropic</td><td><code>claude-sonnet-4-6</code></td><td>200K</td><td>$3.00</td><td>$15.00</td><td>Reasoning, computer use, prompt caching, vision, PDF</td></tr><tr><td>Vertex AI</td><td><code>vertex_ai/claude-opus-4-6@default</code></td><td>1M</td><td>$5.00</td><td>$25.00</td><td>Reasoning, computer use, prompt caching</td></tr><tr><td>Google Gemini</td><td><code>gemini/gemini-3.1-pro-preview</code></td><td>1M</td><td>$2.00</td><td>$12.00</td><td>Audio, video, images, PDF</td></tr><tr><td>Google Gemini</td><td><code>gemini/gemini-3.1-pro-preview-customtools</code></td><td>1M</td><td>$2.00</td><td>$12.00</td><td>Custom tools</td></tr><tr><td>GitHub Copilot</td><td><code>github_copilot/gpt-5.3-codex</code></td><td>128K</td><td>-</td><td>-</td><td>Responses API, function calling, vision</td></tr><tr><td>GitHub Copilot</td><td><code>github_copilot/claude-opus-4.6-fast</code></td><td>128K</td><td>-</td><td>-</td><td>Chat completions, function calling, vision</td></tr><tr><td>Mistral</td><td><code>mistral/devstral-small-latest</code></td><td>256K</td><td>$0.10</td><td>$0.30</td><td>Function calling, response schema</td></tr><tr><td>Mistral</td><td><code>mistral/devstral-latest</code></td><td>256K</td><td>$0.40</td><td>$2.00</td><td>Function calling, response schema</td></tr><tr><td>Mistral</td><td><code>mistral/devstral-medium-latest</code></td><td>256K</td><td>$0.40</td><td>$2.00</td><td>Function calling, response schema</td></tr><tr><td>OpenRouter</td><td><code>openrouter/minimax/minimax-m2.5</code></td><td>196K</td><td>$0.30</td><td>$1.10</td><td>Function calling, reasoning, prompt caching</td></tr><tr><td>Fireworks AI</td><td><code>fireworks_ai/accounts/fireworks/models/glm-4p7</code></td><td>-</td><td>-</td><td>-</td><td>Chat completions</td></tr><tr><td>Fireworks AI</td><td><code>fireworks_ai/accounts/fireworks/models/minimax-m2p1</code></td><td>-</td><td>-</td><td>-</td><td>Chat completions</td></tr><tr><td>Fireworks AI</td><td><code>fireworks_ai/accounts/fireworks/models/kimi-k2p5</code></td><td>-</td><td>-</td><td>-</td><td>Chat completions</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-81-14#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Day 0 support for Claude Sonnet 4.6 with reasoning, computer use, and 200K context - <a href="https://github.com/BerriAI/litellm/pull/21401" target="_blank" rel="noopener noreferrer">PR #21401</a></li>
<li>Add Claude Sonnet 4.6 pricing - <a href="https://github.com/BerriAI/litellm/pull/21395" target="_blank" rel="noopener noreferrer">PR #21395</a></li>
<li>Add day 0 feature support for Claude Sonnet 4.6 (streaming, function calling, vision) - <a href="https://github.com/BerriAI/litellm/pull/21448" target="_blank" rel="noopener noreferrer">PR #21448</a></li>
<li>Add <code>reasoning</code> effort and extended thinking support for Sonnet 4.6 - <a href="https://github.com/BerriAI/litellm/pull/21598" target="_blank" rel="noopener noreferrer">PR #21598</a></li>
<li>Fix empty system messages in <code>translate_system_message</code> - <a href="https://github.com/BerriAI/litellm/pull/21630" target="_blank" rel="noopener noreferrer">PR #21630</a></li>
<li>Sanitize Anthropic messages for multi-turn compatibility - <a href="https://github.com/BerriAI/litellm/pull/21464" target="_blank" rel="noopener noreferrer">PR #21464</a></li>
<li>Map <code>websearch</code> tool from <code>/v1/messages</code> to <code>/chat/completions</code> - <a href="https://github.com/BerriAI/litellm/pull/21465" target="_blank" rel="noopener noreferrer">PR #21465</a></li>
<li>Forward <code>reasoning</code> field as <code>reasoning_content</code> in delta streaming - <a href="https://github.com/BerriAI/litellm/pull/21468" target="_blank" rel="noopener noreferrer">PR #21468</a></li>
<li>Add server-side compaction translation from OpenAI to Anthropic format - <a href="https://github.com/BerriAI/litellm/pull/21555" target="_blank" rel="noopener noreferrer">PR #21555</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">AWS Bedrock</a></strong></p>
<ul>
<li>Native structured outputs API support (<code>outputConfig.textFormat</code>) - <a href="https://github.com/BerriAI/litellm/pull/21222" target="_blank" rel="noopener noreferrer">PR #21222</a></li>
<li>Support <code>nova/</code> and <code>nova-2/</code> spec prefixes for custom imported models - <a href="https://github.com/BerriAI/litellm/pull/21359" target="_blank" rel="noopener noreferrer">PR #21359</a></li>
<li>Broaden Nova 2 model detection to support all <code>nova-2-*</code> variants - <a href="https://github.com/BerriAI/litellm/pull/21358" target="_blank" rel="noopener noreferrer">PR #21358</a></li>
<li>Clamp <code>thinking.budget_tokens</code> to minimum 1024 - <a href="https://github.com/BerriAI/litellm/pull/21306" target="_blank" rel="noopener noreferrer">PR #21306</a></li>
<li>Fix <code>parallel_tool_calls</code> mapping for Bedrock Converse - <a href="https://github.com/BerriAI/litellm/pull/21659" target="_blank" rel="noopener noreferrer">PR #21659</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Google Gemini / Vertex AI</a></strong></p>
<ul>
<li>Day 0 support for <code>gemini-3.1-pro-preview</code> - <a href="https://github.com/BerriAI/litellm/pull/21568" target="_blank" rel="noopener noreferrer">PR #21568</a></li>
<li>Fix <code>_map_reasoning_effort_to_thinking_level</code> for all Gemini 3 family models - <a href="https://github.com/BerriAI/litellm/pull/21654" target="_blank" rel="noopener noreferrer">PR #21654</a></li>
<li>Add reasoning support via config for Gemini models - <a href="https://github.com/BerriAI/litellm/pull/21663" target="_blank" rel="noopener noreferrer">PR #21663</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/databricks">Databricks</a></strong></p>
<ul>
<li>Add Databricks to supported providers for response schema - <a href="https://github.com/BerriAI/litellm/pull/21368" target="_blank" rel="noopener noreferrer">PR #21368</a></li>
<li>Native Responses API support for Databricks GPT models - <a href="https://github.com/BerriAI/litellm/pull/21460" target="_blank" rel="noopener noreferrer">PR #21460</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/github_copilot">GitHub Copilot</a></strong></p>
<ul>
<li>Add <code>github_copilot/gpt-5.3-codex</code> and <code>github_copilot/claude-opus-4.6-fast</code> models - <a href="https://github.com/BerriAI/litellm/pull/21316" target="_blank" rel="noopener noreferrer">PR #21316</a></li>
<li>Fix unsupported params for ChatGPT Codex - <a href="https://github.com/BerriAI/litellm/pull/21209" target="_blank" rel="noopener noreferrer">PR #21209</a></li>
<li>Allow GitHub model aliases to reuse upstream model metadata - <a href="https://github.com/BerriAI/litellm/pull/21497" target="_blank" rel="noopener noreferrer">PR #21497</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/mistral">Mistral</a></strong></p>
<ul>
<li>Add <code>devstral-2512</code> model aliases (<code>devstral-small-latest</code>, <code>devstral-latest</code>, <code>devstral-medium-latest</code>) - <a href="https://github.com/BerriAI/litellm/pull/21372" target="_blank" rel="noopener noreferrer">PR #21372</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/watsonx">IBM watsonx.ai</a></strong></p>
<ul>
<li>Add native rerank support - <a href="https://github.com/BerriAI/litellm/pull/21303" target="_blank" rel="noopener noreferrer">PR #21303</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/xai">xAI</a></strong></p>
<ul>
<li>Fix usage object in xAI responses - <a href="https://github.com/BerriAI/litellm/pull/21559" target="_blank" rel="noopener noreferrer">PR #21559</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/dashscope">Dashscope</a></strong></p>
<ul>
<li>Remove list-to-str transformation that caused incorrect request formatting - <a href="https://github.com/BerriAI/litellm/pull/21547" target="_blank" rel="noopener noreferrer">PR #21547</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vllm">hosted_vllm</a></strong></p>
<ul>
<li>Convert thinking blocks to content blocks for multi-turn conversations - <a href="https://github.com/BerriAI/litellm/pull/21557" target="_blank" rel="noopener noreferrer">PR #21557</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/oci_cohere">OCI / Oracle</a></strong></p>
<ul>
<li>Fix Grok output pricing - <a href="https://github.com/BerriAI/litellm/pull/21329" target="_blank" rel="noopener noreferrer">PR #21329</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">AU Anthropic</a></strong></p>
<ul>
<li>Fix <code>au.anthropic.claude-opus-4-6-v1</code> model ID - <a href="https://github.com/BerriAI/litellm/pull/20731" target="_blank" rel="noopener noreferrer">PR #20731</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Add routing based on reasoning support — skip deployments that don't support reasoning when <code>thinking</code> params are present - <a href="https://github.com/BerriAI/litellm/pull/21302" target="_blank" rel="noopener noreferrer">PR #21302</a></li>
<li>Add <code>stop</code> as supported param for OpenAI and Azure - <a href="https://github.com/BerriAI/litellm/pull/21539" target="_blank" rel="noopener noreferrer">PR #21539</a></li>
<li>Add <code>store</code> and other missing params to <code>OPENAI_CHAT_COMPLETION_PARAMS</code> - <a href="https://github.com/BerriAI/litellm/pull/21195" target="_blank" rel="noopener noreferrer">PR #21195</a>, <a href="https://github.com/BerriAI/litellm/pull/21360" target="_blank" rel="noopener noreferrer">PR #21360</a></li>
<li>Preserve <code>provider_specific_fields</code> from proxy responses - <a href="https://github.com/BerriAI/litellm/pull/21220" target="_blank" rel="noopener noreferrer">PR #21220</a></li>
<li>Add default usage data configuration - <a href="https://github.com/BerriAI/litellm/pull/21550" target="_blank" rel="noopener noreferrer">PR #21550</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-81-14#bug-fixes" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">AWS Bedrock</a></strong></p>
<ul>
<li>Fix service_tier cost propagation - <a href="https://github.com/BerriAI/litellm/pull/21172" target="_blank" rel="noopener noreferrer">PR #21172</a></li>
<li>Fix per-image pricing for multimodal embeddings - <a href="https://github.com/BerriAI/litellm/pull/21646" target="_blank" rel="noopener noreferrer">PR #21646</a></li>
<li>Use <code>batch_</code> prefix for Vertex AI batch IDs in <code>encode_file_id_with_model</code> - <a href="https://github.com/BerriAI/litellm/pull/21624" target="_blank" rel="noopener noreferrer">PR #21624</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock Converse</a></strong></p>
<ul>
<li>Fix Anthropic usage object to match v1/messages spec - <a href="https://github.com/BerriAI/litellm/pull/21295" target="_blank" rel="noopener noreferrer">PR #21295</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/fireworks_ai">Fireworks AI</a></strong></p>
<ul>
<li>Add missing model pricing for <code>glm-4p7</code>, <code>minimax-m2p1</code>, <code>kimi-k2p5</code> - <a href="https://github.com/BerriAI/litellm/pull/21642" target="_blank" rel="noopener noreferrer">PR #21642</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/response_api">Responses API</a></strong></p>
<ul>
<li>Fix <code>use None</code> instead of <code>Reasoning()</code> for reasoning parameter - <a href="https://github.com/BerriAI/litellm/pull/21103" target="_blank" rel="noopener noreferrer">PR #21103</a></li>
<li>Preserve metadata for custom callbacks on codex/responses path - <a href="https://github.com/BerriAI/litellm/pull/21243" target="_blank" rel="noopener noreferrer">PR #21243</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-81-14#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-81-14#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/response_api">Responses API</a></strong></p>
<ul>
<li>Return <code>finish_reason='tool_calls'</code> when response contains function_call items - <a href="https://github.com/BerriAI/litellm/pull/19745" target="_blank" rel="noopener noreferrer">PR #19745</a></li>
<li>Eliminate per-chunk thread spawning in async streaming path for significantly better throughput - <a href="https://github.com/BerriAI/litellm/pull/21709" target="_blank" rel="noopener noreferrer">PR #21709</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/evals_api">Evals API</a></strong></p>
<ul>
<li>Add support for OpenAI Evals API - <a href="https://github.com/BerriAI/litellm/pull/21375" target="_blank" rel="noopener noreferrer">PR #21375</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/batches">Batch API</a></strong></p>
<ul>
<li>Add file deletion criteria with batch references - <a href="https://github.com/BerriAI/litellm/pull/21456" target="_blank" rel="noopener noreferrer">PR #21456</a></li>
<li>Misc bug fixes for managed batches - <a href="https://github.com/BerriAI/litellm/pull/21157" target="_blank" rel="noopener noreferrer">PR #21157</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/pass_through/bedrock">Pass-Through Endpoints</a></strong></p>
<ul>
<li>Add method-based routing for passthrough endpoints - <a href="https://github.com/BerriAI/litellm/pull/21543" target="_blank" rel="noopener noreferrer">PR #21543</a></li>
<li>Preserve and forward OAuth Authorization headers through proxy layer - <a href="https://github.com/BerriAI/litellm/pull/19912" target="_blank" rel="noopener noreferrer">PR #19912</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/completion/input">Websearch / Tool Calling</a></strong></p>
<ul>
<li>Add DuckDuckGo as a search tool - <a href="https://github.com/BerriAI/litellm/pull/21467" target="_blank" rel="noopener noreferrer">PR #21467</a></li>
<li>Fix <code>pre_call_deployment_hook</code> not triggering via proxy router for websearch - <a href="https://github.com/BerriAI/litellm/pull/21433" target="_blank" rel="noopener noreferrer">PR #21433</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Exclude tool params for models without function calling support - <a href="https://github.com/BerriAI/litellm/pull/21244" target="_blank" rel="noopener noreferrer">PR #21244</a></li>
<li>Add <code>store</code> param to OpenAI chat completion params - <a href="https://github.com/BerriAI/litellm/pull/21195" target="_blank" rel="noopener noreferrer">PR #21195</a></li>
<li>Add reasoning support via config for per-model reasoning configuration - <a href="https://github.com/BerriAI/litellm/pull/21663" target="_blank" rel="noopener noreferrer">PR #21663</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs">Bugs<a href="https://docs.litellm.ai/release_notes/v1-81-14#bugs" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>General</strong>
<ul>
<li>Fix <code>api_base</code> resolution error for models with multiple potential endpoints - <a href="https://github.com/BerriAI/litellm/pull/21658" target="_blank" rel="noopener noreferrer">PR #21658</a></li>
<li>Fix session grouping broken for dict rows from <code>query_raw</code> - <a href="https://github.com/BerriAI/litellm/pull/21435" target="_blank" rel="noopener noreferrer">PR #21435</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-81-14#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-81-14#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong>Access Groups</strong></p>
<ul>
<li>Add Access Group Selector to Create and Edit flow for Keys/Teams - <a href="https://github.com/BerriAI/litellm/pull/21234" target="_blank" rel="noopener noreferrer">PR #21234</a></li>
</ul>
</li>
<li>
<p><strong>Virtual Keys</strong></p>
<ul>
<li>Fix virtual key grace period from env/UI - <a href="https://github.com/BerriAI/litellm/pull/20321" target="_blank" rel="noopener noreferrer">PR #20321</a></li>
<li>Fix key expiry default duration - <a href="https://github.com/BerriAI/litellm/pull/21362" target="_blank" rel="noopener noreferrer">PR #21362</a></li>
<li>Key Last Active Tracking — see when a key was last used - <a href="https://github.com/BerriAI/litellm/pull/21545" target="_blank" rel="noopener noreferrer">PR #21545</a></li>
<li>Fix <code>/v1/models</code> returning wildcard instead of expanded models for BYOK team keys - <a href="https://github.com/BerriAI/litellm/pull/21408" target="_blank" rel="noopener noreferrer">PR #21408</a></li>
<li>Return <code>failed_tokens</code> in delete_verification_tokens response - <a href="https://github.com/BerriAI/litellm/pull/21609" target="_blank" rel="noopener noreferrer">PR #21609</a></li>
</ul>
</li>
<li>
<p><strong>Models + Endpoints</strong></p>
<ul>
<li>Add Model Settings Modal to Models &amp; Endpoints page - <a href="https://github.com/BerriAI/litellm/pull/21516" target="_blank" rel="noopener noreferrer">PR #21516</a></li>
<li>Allow <code>store_model_in_db</code> to be set via database (not just config) - <a href="https://github.com/BerriAI/litellm/pull/21511" target="_blank" rel="noopener noreferrer">PR #21511</a></li>
<li>Fix <code>input_cost_per_token</code> masked/hidden in Model Info UI - <a href="https://github.com/BerriAI/litellm/pull/21723" target="_blank" rel="noopener noreferrer">PR #21723</a></li>
<li>Fix credentials for UI-created models in batch file uploads - <a href="https://github.com/BerriAI/litellm/pull/21502" target="_blank" rel="noopener noreferrer">PR #21502</a></li>
<li>Resolve credentials for UI-created models - <a href="https://github.com/BerriAI/litellm/pull/21502" target="_blank" rel="noopener noreferrer">PR #21502</a></li>
</ul>
</li>
<li>
<p><strong>Teams</strong></p>
<ul>
<li>Allow team members to view entire team usage - <a href="https://github.com/BerriAI/litellm/pull/21537" target="_blank" rel="noopener noreferrer">PR #21537</a></li>
<li>Fix service account visibility for team members - <a href="https://github.com/BerriAI/litellm/pull/21627" target="_blank" rel="noopener noreferrer">PR #21627</a></li>
<li>Organization Info page: show member email, AntD tabs, reusable MemberTable - <a href="https://github.com/BerriAI/litellm/pull/21745" target="_blank" rel="noopener noreferrer">PR #21745</a></li>
</ul>
</li>
<li>
<p><strong>Usage / Spend Logs</strong></p>
<ul>
<li>Allow filtering Usage by User - <a href="https://github.com/BerriAI/litellm/pull/21351" target="_blank" rel="noopener noreferrer">PR #21351</a></li>
<li>Inject Credential Name as Tag for Usage Page filtering - <a href="https://github.com/BerriAI/litellm/pull/21715" target="_blank" rel="noopener noreferrer">PR #21715</a></li>
<li>Prefix credential tags and update Tag usage banner - <a href="https://github.com/BerriAI/litellm/pull/21739" target="_blank" rel="noopener noreferrer">PR #21739</a></li>
<li>Show retry count for requests in Logs view - <a href="https://github.com/BerriAI/litellm/pull/21704" target="_blank" rel="noopener noreferrer">PR #21704</a></li>
<li>Fix Aggregated Daily Activity Endpoint performance - <a href="https://github.com/BerriAI/litellm/pull/21613" target="_blank" rel="noopener noreferrer">PR #21613</a></li>
</ul>
</li>
<li>
<p><strong>SSO / Auth</strong></p>
<ul>
<li>Fix SSO PKCE support in multi-pod Kubernetes deployments - <a href="https://github.com/BerriAI/litellm/pull/20314" target="_blank" rel="noopener noreferrer">PR #20314</a></li>
<li>Preserve SSO role regardless of <code>role_mappings</code> config - <a href="https://github.com/BerriAI/litellm/pull/21503" target="_blank" rel="noopener noreferrer">PR #21503</a></li>
</ul>
</li>
<li>
<p><strong>Proxy CLI / Master Key</strong></p>
<ul>
<li>Fix master key rotation Prisma validation errors - <a href="https://github.com/BerriAI/litellm/pull/21330" target="_blank" rel="noopener noreferrer">PR #21330</a></li>
<li>Handle missing <code>DATABASE_URL</code> in <code>append_query_params</code> - <a href="https://github.com/BerriAI/litellm/pull/21239" target="_blank" rel="noopener noreferrer">PR #21239</a></li>
</ul>
</li>
<li>
<p><strong>Project Management</strong></p>
<ul>
<li>Add Project Management APIs for organizing resources - <a href="https://github.com/BerriAI/litellm/pull/21078" target="_blank" rel="noopener noreferrer">PR #21078</a></li>
</ul>
</li>
<li>
<p><strong>UI Improvements</strong></p>
<ul>
<li>Content Filters: help edit/view categories and 1-click add with pagination - <a href="https://github.com/BerriAI/litellm/pull/21223" target="_blank" rel="noopener noreferrer">PR #21223</a></li>
<li>Playground: test fallbacks with UI - <a href="https://github.com/BerriAI/litellm/pull/21007" target="_blank" rel="noopener noreferrer">PR #21007</a></li>
<li>Add <code>forward_client_headers_to_llm_api</code> toggle to general settings - <a href="https://github.com/BerriAI/litellm/pull/21776" target="_blank" rel="noopener noreferrer">PR #21776</a></li>
<li>Fix <code>is_premium()</code> debug log spam on every request - <a href="https://github.com/BerriAI/litellm/pull/20841" target="_blank" rel="noopener noreferrer">PR #20841</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs-1">Bugs<a href="https://docs.litellm.ai/release_notes/v1-81-14#bugs-1" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li>Spend Logs: Fix cost calculation - <a href="https://github.com/BerriAI/litellm/pull/21152" target="_blank" rel="noopener noreferrer">PR #21152</a></li>
<li>Logs: Fix table not updating and pagination issues - <a href="https://github.com/BerriAI/litellm/pull/21708" target="_blank" rel="noopener noreferrer">PR #21708</a></li>
<li>Fix <code>/get_image</code> ignoring <code>UI_LOGO_PATH</code> when <code>cached_logo.jpg</code> exists - <a href="https://github.com/BerriAI/litellm/pull/21637" target="_blank" rel="noopener noreferrer">PR #21637</a></li>
<li>Fix duplicate URL in <code>tagsSpendLogsCall</code> query string - <a href="https://github.com/BerriAI/litellm/pull/20909" target="_blank" rel="noopener noreferrer">PR #20909</a></li>
<li>Preserve <code>key_alias</code> and <code>team_id</code> metadata in <code>/user/daily/activity/aggregated</code> after key deletion or regeneration - <a href="https://github.com/BerriAI/litellm/pull/20684" target="_blank" rel="noopener noreferrer">PR #20684</a></li>
<li>Uncomment <code>response_model</code> in <code>user_info</code> endpoint - <a href="https://github.com/BerriAI/litellm/pull/17430" target="_blank" rel="noopener noreferrer">PR #17430</a></li>
<li>Allow <code>internal_user_viewer</code> to access RAG endpoints; restrict ingest to existing vector stores - <a href="https://github.com/BerriAI/litellm/pull/21508" target="_blank" rel="noopener noreferrer">PR #21508</a></li>
<li>Suppress warning for <code>litellm-dashboard</code> team in agent permission handler - <a href="https://github.com/BerriAI/litellm/pull/21721" target="_blank" rel="noopener noreferrer">PR #21721</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="ai-integrations">AI Integrations<a href="https://docs.litellm.ai/release_notes/v1-81-14#ai-integrations" class="hash-link" aria-label="Direct link to AI Integrations" title="Direct link to AI Integrations">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="logging">Logging<a href="https://docs.litellm.ai/release_notes/v1-81-14#logging" class="hash-link" aria-label="Direct link to Logging" title="Direct link to Logging">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#datadog">DataDog</a></strong></p>
<ul>
<li>Add <code>team</code> tag to logs, metrics, and cost management - <a href="https://github.com/BerriAI/litellm/pull/21449" target="_blank" rel="noopener noreferrer">PR #21449</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#prometheus">Prometheus</a></strong></p>
<ul>
<li>Fix double-counting of <code>litellm_proxy_total_requests_metric</code> - <a href="https://github.com/BerriAI/litellm/pull/21159" target="_blank" rel="noopener noreferrer">PR #21159</a></li>
<li>Guard against None metadata in Prometheus metrics - <a href="https://github.com/BerriAI/litellm/pull/21489" target="_blank" rel="noopener noreferrer">PR #21489</a></li>
<li>Add ASGI middleware for improved Prometheus metrics collection - <a href="https://github.com/BerriAI/litellm/pull/20434" target="_blank" rel="noopener noreferrer">PR #20434</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#langfuse">Langfuse</a></strong></p>
<ul>
<li>Improve Langfuse test isolation (multiple stability fixes) - <a href="https://github.com/BerriAI/litellm/pull/21214" target="_blank" rel="noopener noreferrer">PR #21214</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Fix cost to 0 for cached responses in logging - <a href="https://github.com/BerriAI/litellm/pull/21816" target="_blank" rel="noopener noreferrer">PR #21816</a></li>
<li>Improve streaming proxy throughput by fixing middleware and logging bottlenecks - <a href="https://github.com/BerriAI/litellm/pull/21501" target="_blank" rel="noopener noreferrer">PR #21501</a></li>
<li>Reduce proxy overhead for large base64 payloads - <a href="https://github.com/BerriAI/litellm/pull/21594" target="_blank" rel="noopener noreferrer">PR #21594</a></li>
<li>Close streaming connections to prevent connection pool exhaustion - <a href="https://github.com/BerriAI/litellm/pull/21213" target="_blank" rel="noopener noreferrer">PR #21213</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails">Guardrails<a href="https://docs.litellm.ai/release_notes/v1-81-14#guardrails" class="hash-link" aria-label="Direct link to Guardrails" title="Direct link to Guardrails">​</a></h3>
<ul>
<li>
<p><strong>Guardrail Garden</strong></p>
<ul>
<li>Launch Guardrail Garden — a marketplace for pre-built guardrails deployable in one click - <a href="https://github.com/BerriAI/litellm/pull/21732" target="_blank" rel="noopener noreferrer">PR #21732</a></li>
<li>Redesign guardrail creation form with vertical stepper UI - <a href="https://github.com/BerriAI/litellm/pull/21727" target="_blank" rel="noopener noreferrer">PR #21727</a></li>
<li>Add guardrail jump link in log detail view - <a href="https://github.com/BerriAI/litellm/pull/21437" target="_blank" rel="noopener noreferrer">PR #21437</a></li>
<li>Guardrail tracing UI: show policy, detection method, and match details - <a href="https://github.com/BerriAI/litellm/pull/21349" target="_blank" rel="noopener noreferrer">PR #21349</a></li>
</ul>
</li>
<li>
<p><strong>AI Policy Templates</strong></p>
<ul>
<li>Seven new ready-to-deploy policy templates ship in this release:<!-- -->
<ul>
<li>GDPR Art. 32 EU PII Protection - <a href="https://github.com/BerriAI/litellm/pull/21340" target="_blank" rel="noopener noreferrer">PR #21340</a></li>
<li>EU AI Act Article 5 (5 sub-guardrails, with French language support) - <a href="https://github.com/BerriAI/litellm/pull/21342" target="_blank" rel="noopener noreferrer">PR #21342</a>, <a href="https://github.com/BerriAI/litellm/pull/21453" target="_blank" rel="noopener noreferrer">PR #21453</a>, <a href="https://github.com/BerriAI/litellm/pull/21427" target="_blank" rel="noopener noreferrer">PR #21427</a></li>
<li>Prompt injection detection - <a href="https://github.com/BerriAI/litellm/pull/21520" target="_blank" rel="noopener noreferrer">PR #21520</a></li>
<li>Aviation and UAE topic filters with tag-based routing - <a href="https://github.com/BerriAI/litellm/pull/21518" target="_blank" rel="noopener noreferrer">PR #21518</a></li>
<li>Airline off-topic restriction - <a href="https://github.com/BerriAI/litellm/pull/21607" target="_blank" rel="noopener noreferrer">PR #21607</a></li>
<li>SQL injection - <a href="https://github.com/BerriAI/litellm/pull/21806" target="_blank" rel="noopener noreferrer">PR #21806</a></li>
</ul>
</li>
<li>AI-powered policy template suggestions with latency overhead estimates - <a href="https://github.com/BerriAI/litellm/pull/21589" target="_blank" rel="noopener noreferrer">PR #21589</a>, <a href="https://github.com/BerriAI/litellm/pull/21608" target="_blank" rel="noopener noreferrer">PR #21608</a>, <a href="https://github.com/BerriAI/litellm/pull/21620" target="_blank" rel="noopener noreferrer">PR #21620</a></li>
</ul>
</li>
<li>
<p><strong>Compliance Checker</strong></p>
<ul>
<li>Add compliance checker endpoints + UI panel - <a href="https://github.com/BerriAI/litellm/pull/21432" target="_blank" rel="noopener noreferrer">PR #21432</a></li>
<li>CSV dataset upload to compliance playground for batch testing - <a href="https://github.com/BerriAI/litellm/pull/21526" target="_blank" rel="noopener noreferrer">PR #21526</a></li>
</ul>
</li>
<li>
<p><strong>Built-in Guardrails</strong></p>
<ul>
<li>Competitor name blocker: blocks by name, handles streaming, supports name variations, and splits pre/post call - <a href="https://github.com/BerriAI/litellm/pull/21719" target="_blank" rel="noopener noreferrer">PR #21719</a>, <a href="https://github.com/BerriAI/litellm/pull/21533" target="_blank" rel="noopener noreferrer">PR #21533</a></li>
<li>Topic blocker with both keyword and embedding-based implementations - <a href="https://github.com/BerriAI/litellm/pull/21713" target="_blank" rel="noopener noreferrer">PR #21713</a></li>
<li>Insults content filter - <a href="https://github.com/BerriAI/litellm/pull/21729" target="_blank" rel="noopener noreferrer">PR #21729</a></li>
<li>MCP Security guardrail to block unregistered MCP servers - <a href="https://github.com/BerriAI/litellm/pull/21429" target="_blank" rel="noopener noreferrer">PR #21429</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">Generic Guardrails</a></strong></p>
<ul>
<li>Add configurable fallback to handle generic guardrail endpoint connection failures - <a href="https://github.com/BerriAI/litellm/pull/21245" target="_blank" rel="noopener noreferrer">PR #21245</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">Presidio</a></strong></p>
<ul>
<li>Fix Presidio controls configuration - <a href="https://github.com/BerriAI/litellm/pull/21798" target="_blank" rel="noopener noreferrer">PR #21798</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">LakeraAI</a></strong></p>
<ul>
<li>Avoid <code>KeyError</code> on missing <code>LAKERA_API_KEY</code> during initialization - <a href="https://github.com/BerriAI/litellm/pull/21422" target="_blank" rel="noopener noreferrer">PR #21422</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="auto-routing">Auto Routing<a href="https://docs.litellm.ai/release_notes/v1-81-14#auto-routing" class="hash-link" aria-label="Direct link to Auto Routing" title="Direct link to Auto Routing">​</a></h3>
<ul>
<li><strong>Complexity-based auto routing</strong> — new router strategy that scores requests across 7 dimensions (token count, code presence, reasoning markers, technical terms, etc.) and routes to the appropriate model tier — no embeddings or API calls required - <a href="https://github.com/BerriAI/litellm/pull/21789" target="_blank" rel="noopener noreferrer">PR #21789</a>, <a href="https://docs.litellm.ai/docs/proxy/auto_routing">Docs</a></li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="prompt-management">Prompt Management<a href="https://docs.litellm.ai/release_notes/v1-81-14#prompt-management" class="hash-link" aria-label="Direct link to Prompt Management" title="Direct link to Prompt Management">​</a></h3>
<ul>
<li><strong>Prompt Management API</strong>
<ul>
<li>New API to interact with prompt management integrations without requiring a PR - <a href="https://github.com/BerriAI/litellm/pull/17800" target="_blank" rel="noopener noreferrer">PR #17800</a>, <a href="https://github.com/BerriAI/litellm/pull/17946" target="_blank" rel="noopener noreferrer">PR #17946</a></li>
<li>Fix prompt registry configuration issues - <a href="https://github.com/BerriAI/litellm/pull/21402" target="_blank" rel="noopener noreferrer">PR #21402</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-81-14#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li><strong>Fix Bedrock service_tier cost propagation</strong> — costs from service-tier responses now correctly flow through to spend tracking - <a href="https://github.com/BerriAI/litellm/pull/21172" target="_blank" rel="noopener noreferrer">PR #21172</a></li>
<li><strong>Fix cost for cached responses</strong> — cached responses now correctly log $0 cost instead of re-billing - <a href="https://github.com/BerriAI/litellm/pull/21816" target="_blank" rel="noopener noreferrer">PR #21816</a></li>
<li><strong>Aggregate daily activity endpoint performance</strong> — faster queries for <code>/user/daily/activity/aggregated</code> - <a href="https://github.com/BerriAI/litellm/pull/21613" target="_blank" rel="noopener noreferrer">PR #21613</a></li>
<li><strong>Preserve key_alias and team_id metadata</strong> in <code>/user/daily/activity/aggregated</code> after key deletion or regeneration - <a href="https://github.com/BerriAI/litellm/pull/20684" target="_blank" rel="noopener noreferrer">PR #20684</a></li>
<li><strong>Inject Credential Name as Tag</strong> for granular usage page filtering by credential - <a href="https://github.com/BerriAI/litellm/pull/21715" target="_blank" rel="noopener noreferrer">PR #21715</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-gateway">MCP Gateway<a href="https://docs.litellm.ai/release_notes/v1-81-14#mcp-gateway" class="hash-link" aria-label="Direct link to MCP Gateway" title="Direct link to MCP Gateway">​</a></h2>
<ul>
<li><strong>OpenAPI-to-MCP</strong> — Convert any OpenAPI spec to an MCP server via API or UI - <a href="https://github.com/BerriAI/litellm/pull/21575" target="_blank" rel="noopener noreferrer">PR #21575</a>, <a href="https://github.com/BerriAI/litellm/pull/21662" target="_blank" rel="noopener noreferrer">PR #21662</a></li>
<li><strong>MCP User Permissions</strong> — Fine-grained permissions for end users on MCP servers - <a href="https://github.com/BerriAI/litellm/pull/21462" target="_blank" rel="noopener noreferrer">PR #21462</a></li>
<li><strong>MCP Security Guardrail</strong> — Block calls to unregistered MCP servers - <a href="https://github.com/BerriAI/litellm/pull/21429" target="_blank" rel="noopener noreferrer">PR #21429</a></li>
<li><strong>Fix StreamableHTTPSessionManager</strong> — Revert to stateless mode to prevent session state issues - <a href="https://github.com/BerriAI/litellm/pull/21323" target="_blank" rel="noopener noreferrer">PR #21323</a></li>
<li><strong>Fix Bedrock AgentCore Accept header</strong> — Add required Accept header for AgentCore MCP server requests - <a href="https://github.com/BerriAI/litellm/pull/21551" target="_blank" rel="noopener noreferrer">PR #21551</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements">Performance / Loadbalancing / Reliability improvements<a href="https://docs.litellm.ai/release_notes/v1-81-14#performance--loadbalancing--reliability-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements" title="Direct link to Performance / Loadbalancing / Reliability improvements">​</a></h2>
<p><strong>Logging &amp; callback overhead</strong></p>
<ul>
<li>Move async/sync callback separation from per-request to callback registration time — ~30% speedup for callback-heavy deployments - <a href="https://github.com/BerriAI/litellm/pull/20354" target="_blank" rel="noopener noreferrer">PR #20354</a></li>
<li>Skip Pydantic Usage round-trip in logging payload — reduces serialization overhead per request - <a href="https://github.com/BerriAI/litellm/pull/21003" target="_blank" rel="noopener noreferrer">PR #21003</a></li>
<li>Skip duplicate <code>get_standard_logging_object_payload</code> calls for non-streaming requests - <a href="https://github.com/BerriAI/litellm/pull/20440" target="_blank" rel="noopener noreferrer">PR #20440</a></li>
<li>Reuse <code>LiteLLM_Params</code> object across the request lifecycle - <a href="https://github.com/BerriAI/litellm/pull/20593" target="_blank" rel="noopener noreferrer">PR #20593</a></li>
<li>Optimize <code>add_litellm_data_to_request</code> hot path - <a href="https://github.com/BerriAI/litellm/pull/20526" target="_blank" rel="noopener noreferrer">PR #20526</a></li>
<li>Optimize <code>model_dump_with_preserved_fields</code> - <a href="https://github.com/BerriAI/litellm/pull/20882" target="_blank" rel="noopener noreferrer">PR #20882</a></li>
<li>Pre-compute OpenAI client init params at module load instead of per-request - <a href="https://github.com/BerriAI/litellm/pull/20789" target="_blank" rel="noopener noreferrer">PR #20789</a></li>
<li>Reduce proxy overhead for large base64 payloads - <a href="https://github.com/BerriAI/litellm/pull/21594" target="_blank" rel="noopener noreferrer">PR #21594</a></li>
<li>Improve streaming proxy throughput by fixing middleware and logging bottlenecks - <a href="https://github.com/BerriAI/litellm/pull/21501" target="_blank" rel="noopener noreferrer">PR #21501</a></li>
<li>Eliminate per-chunk thread spawning in Responses API async streaming - <a href="https://github.com/BerriAI/litellm/pull/21709" target="_blank" rel="noopener noreferrer">PR #21709</a></li>
</ul>
<p><strong>Cost calculation</strong></p>
<ul>
<li>Optimize <code>completion_cost()</code> with early-exit and caching - <a href="https://github.com/BerriAI/litellm/pull/20448" target="_blank" rel="noopener noreferrer">PR #20448</a></li>
<li>Cost calculator: reduce repeated lookups and dict copies - <a href="https://github.com/BerriAI/litellm/pull/20541" target="_blank" rel="noopener noreferrer">PR #20541</a></li>
</ul>
<p><strong>Router &amp; load balancing</strong></p>
<ul>
<li>Remove quadratic deployment scan in usage-based routing v2 - <a href="https://github.com/BerriAI/litellm/pull/21211" target="_blank" rel="noopener noreferrer">PR #21211</a></li>
<li>Avoid O(n²) membership scans in team deployment filter - <a href="https://github.com/BerriAI/litellm/pull/21210" target="_blank" rel="noopener noreferrer">PR #21210</a></li>
<li>Avoid O(n) alias scan for non-alias <code>get_model_list</code> lookups - <a href="https://github.com/BerriAI/litellm/pull/21136" target="_blank" rel="noopener noreferrer">PR #21136</a></li>
<li>Increase default LRU cache size to reduce multi-model cache thrash - <a href="https://github.com/BerriAI/litellm/pull/21139" target="_blank" rel="noopener noreferrer">PR #21139</a></li>
<li>Cache <code>get_model_access_groups()</code> no-args result on Router - <a href="https://github.com/BerriAI/litellm/pull/20374" target="_blank" rel="noopener noreferrer">PR #20374</a></li>
<li>Deployment affinity routing callback — route to the same deployment for a session - <a href="https://github.com/BerriAI/litellm/pull/19143" target="_blank" rel="noopener noreferrer">PR #19143</a></li>
<li>Session-ID-based routing — use <code>session_id</code> for consistent routing within a session - <a href="https://github.com/BerriAI/litellm/pull/21763" target="_blank" rel="noopener noreferrer">PR #21763</a></li>
</ul>
<p><strong>Connection management &amp; reliability</strong></p>
<ul>
<li>Fix Redis connection pool reliability — prevent connection exhaustion under load - <a href="https://github.com/BerriAI/litellm/pull/21717" target="_blank" rel="noopener noreferrer">PR #21717</a></li>
<li>Fix Prisma connection self-heal for auth and runtime reconnection (reverted, will be re-introduced with fixes) - <a href="https://github.com/BerriAI/litellm/pull/21706" target="_blank" rel="noopener noreferrer">PR #21706</a></li>
<li>Close streaming connections to prevent connection pool exhaustion - <a href="https://github.com/BerriAI/litellm/pull/21213" target="_blank" rel="noopener noreferrer">PR #21213</a></li>
<li>Make <code>PodLockManager.release_lock</code> atomic compare-and-delete - <a href="https://github.com/BerriAI/litellm/pull/21226" target="_blank" rel="noopener noreferrer">PR #21226</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="database-changes">Database Changes<a href="https://docs.litellm.ai/release_notes/v1-81-14#database-changes" class="hash-link" aria-label="Direct link to Database Changes" title="Direct link to Database Changes">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="schema-updates">Schema Updates<a href="https://docs.litellm.ai/release_notes/v1-81-14#schema-updates" class="hash-link" aria-label="Direct link to Schema Updates" title="Direct link to Schema Updates">​</a></h3>
<table><thead><tr><th>Table</th><th>Change Type</th><th>Description</th><th>PR</th></tr></thead><tbody><tr><td><code>LiteLLM_DeletedVerificationToken</code></td><td>New Column</td><td>Added <code>project_id</code> column</td><td><a href="https://github.com/BerriAI/litellm/pull/21587" target="_blank" rel="noopener noreferrer">PR #21587</a></td></tr><tr><td><code>LiteLLM_ProjectTable</code></td><td>New Table</td><td>Project management for organizing resources</td><td><a href="https://github.com/BerriAI/litellm/pull/21078" target="_blank" rel="noopener noreferrer">PR #21078</a></td></tr><tr><td><code>LiteLLM_VerificationToken</code></td><td>New Column</td><td>Added <code>last_active</code> timestamp for key activity tracking</td><td><a href="https://github.com/BerriAI/litellm/pull/21545" target="_blank" rel="noopener noreferrer">PR #21545</a></td></tr><tr><td><code>LiteLLM_ManagedVectorStoreTable</code></td><td>Migration</td><td>Make vector store migration idempotent</td><td><a href="https://github.com/BerriAI/litellm/pull/21325" target="_blank" rel="noopener noreferrer">PR #21325</a></td></tr></tbody></table>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="security">Security<a href="https://docs.litellm.ai/release_notes/v1-81-14#security" class="hash-link" aria-label="Direct link to Security" title="Direct link to Security">​</a></h2>
<p>We run <a href="https://github.com/anchore/grype" target="_blank" rel="noopener noreferrer">Grype</a> and <a href="https://github.com/aquasecurity/trivy" target="_blank" rel="noopener noreferrer">Trivy</a> security scans on every LiteLLM Docker image. Here's the vulnerability report for this release across all published images:</p>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="docker-image-scan-summary">Docker Image Scan Summary<a href="https://docs.litellm.ai/release_notes/v1-81-14#docker-image-scan-summary" class="hash-link" aria-label="Direct link to Docker Image Scan Summary" title="Direct link to Docker Image Scan Summary">​</a></h3>
<table><thead><tr><th>Image</th><th>Critical</th><th>High</th><th>Medium</th><th>Low</th></tr></thead><tbody><tr><td><code>ghcr.io/berriai/litellm:main-latest</code></td><td><strong>0</strong> ✅</td><td>4 unique CVEs</td><td>4</td><td>1</td></tr><tr><td><code>ghcr.io/berriai/litellm-ee:main-latest</code></td><td><strong>0</strong> ✅</td><td>4 unique CVEs</td><td>4</td><td>1</td></tr><tr><td><code>ghcr.io/berriai/litellm-non_root:main-latest</code></td><td><strong>1</strong></td><td>11 unique CVEs</td><td>6</td><td>2</td></tr><tr><td><code>ghcr.io/berriai/litellm-database:main-latest</code></td><td><strong>1</strong></td><td>7 unique CVEs</td><td>5</td><td>1</td></tr><tr><td><code>ghcr.io/berriai/litellm-spend_logs:main-latest</code></td><td><strong>4</strong></td><td>35 matches</td><td>40</td><td>10</td></tr></tbody></table>
<div class="theme-admonition theme-admonition-note admonition_xJq3 alert alert--secondary"><div class="admonitionHeading_Gvgb"><span class="admonitionIcon_Rf37"><svg viewBox="0 0 14 16"><path fill-rule="evenodd" d="M6.3 5.69a.942.942 0 0 1-.28-.7c0-.28.09-.52.28-.7.19-.18.42-.28.7-.28.28 0 .52.09.7.28.18.19.28.42.28.7 0 .28-.09.52-.28.7a1 1 0 0 1-.7.3c-.28 0-.52-.11-.7-.3zM8 7.99c-.02-.25-.11-.48-.31-.69-.2-.19-.42-.3-.69-.31H6c-.27.02-.48.13-.69.31-.2.2-.3.44-.31.69h1v3c.02.27.11.5.31.69.2.2.42.31.69.31h1c.27 0 .48-.11.69-.31.2-.19.3-.42.31-.69H8V7.98v.01zM7 2.3c-3.14 0-5.7 2.54-5.7 5.68 0 3.14 2.56 5.7 5.7 5.7s5.7-2.55 5.7-5.7c0-3.15-2.56-5.69-5.7-5.69v.01zM7 .98c3.86 0 7 3.14 7 7s-3.14 7-7 7-7-3.12-7-7 3.14-7 7-7z"></path></svg></span>note</div><div class="admonitionContent_BuS1"><p>Vulnerability counts are based on full image scans including build-time tooling. High match counts are often inflated by packages like <code>minimatch</code> appearing at multiple versions; the unique CVE counts above reflect the actual distinct vulnerabilities.</p></div></div>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="critical-severity">Critical Severity<a href="https://docs.litellm.ai/release_notes/v1-81-14#critical-severity" class="hash-link" aria-label="Direct link to Critical Severity" title="Direct link to Critical Severity">​</a></h3>
<p><strong>1. Node.js Critical (non-root, database, spend_logs images):</strong>
Node.js 24.12.0 is used <strong>only</strong> for the Admin UI build and Prisma client generation — it is <strong>not</strong> part of the LiteLLM Python application runtime.</p>
<table><thead><tr><th>Package</th><th>Vulnerability</th><th>Description</th><th>Fix Version</th></tr></thead><tbody><tr><td><code>node</code></td><td>CVE-2025-55130</td><td>Node.js critical vulnerability</td><td>20.20.0</td></tr></tbody></table>
<p><strong>2. OpenSSL &amp; Go Critical (spend_logs image only):</strong>
The <code>spend_logs</code> image contains additional vulnerabilities in the underlying Go modules and system libraries.</p>
<table><thead><tr><th>Package</th><th>Vulnerability</th><th>Description</th><th>Fix Version</th></tr></thead><tbody><tr><td><code>libcrypto3</code>, <code>libssl3</code></td><td>CVE-2025-15467</td><td>OpenSSL critical vulnerability</td><td>3.3.6-r0</td></tr><tr><td><code>stdlib</code> (Go)</td><td>CVE-2025-68121</td><td>Go standard library critical vulnerability</td><td>1.24.13+</td></tr></tbody></table>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="high-severity">High Severity<a href="https://docs.litellm.ai/release_notes/v1-81-14#high-severity" class="hash-link" aria-label="Direct link to High Severity" title="Direct link to High Severity">​</a></h3>
<p>All high-severity vulnerabilities are in <strong>npm/Node.js build-time dependencies</strong> or system-level libraries — they are <strong>not</strong> in the LiteLLM Python application code.</p>
<p><strong>Present in all images:</strong></p>
<table><thead><tr><th>Package</th><th>Vulnerability</th><th>Description</th><th>Fix Version</th></tr></thead><tbody><tr><td><code>minimatch</code></td><td>CVE-2026-26996</td><td>DoS via specially crafted glob patterns</td><td>10.2.1+ / 9.0.6+</td></tr><tr><td><code>minimatch</code></td><td>CVE-2026-27903</td><td>DoS due to unbounded recursive backtracking</td><td>10.2.3+ / 9.0.7+</td></tr><tr><td><code>minimatch</code></td><td>CVE-2026-27904</td><td>DoS via catastrophic backtracking in glob expressions</td><td>10.2.3+ / 9.0.7+</td></tr><tr><td><code>tar</code></td><td>CVE-2026-26960 / GHSA-83g3-92jg-28cx</td><td>Arbitrary file read/write via malicious archive hardlinks</td><td>7.5.8</td></tr></tbody></table>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="medium-severity-all-images">Medium Severity (all images)<a href="https://docs.litellm.ai/release_notes/v1-81-14#medium-severity-all-images" class="hash-link" aria-label="Direct link to Medium Severity (all images)" title="Direct link to Medium Severity (all images)">​</a></h3>
<table><thead><tr><th>Package</th><th>Vulnerability</th><th>Status</th></tr></thead><tbody><tr><td><code>pypdf</code> 6.7.2</td><td>GHSA-x7hp-r3qg-r3cj</td><td>Fix available in 6.7.3</td></tr><tr><td>Python 3.13</td><td>CVE-2025-15366, CVE-2025-15367, CVE-2025-12781</td><td>No upstream fix available</td></tr></tbody></table>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="recommendations">Recommendations<a href="https://docs.litellm.ai/release_notes/v1-81-14#recommendations" class="hash-link" aria-label="Direct link to Recommendations" title="Direct link to Recommendations">​</a></h3>
<ul>
<li><strong>LiteLLM Main &amp; EE images</strong> (<code>litellm:main-latest</code>, <code>litellm-ee:main-latest</code>) have the best security posture with <strong>0 critical vulnerabilities</strong>.</li>
<li>All HIGH/CRITICAL findings in the main images relate to build-time Node.js/npm tooling, not the Python runtime.</li>
<li>We are actively monitoring upstream Python and system library fixes for remaining medium-severity vulnerabilities.</li>
</ul>
<p>To report a security vulnerability, email <a href="mailto:support@berri.ai" target="_blank" rel="noopener noreferrer">support@berri.ai</a> with details and steps to reproduce.</p>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates">Documentation Updates<a href="https://docs.litellm.ai/release_notes/v1-81-14#documentation-updates" class="hash-link" aria-label="Direct link to Documentation Updates" title="Direct link to Documentation Updates">​</a></h2>
<ul>
<li>Add OpenAI Agents SDK with LiteLLM guide - <a href="https://github.com/BerriAI/litellm/pull/21311" target="_blank" rel="noopener noreferrer">PR #21311</a></li>
<li>Access Groups documentation - <a href="https://github.com/BerriAI/litellm/pull/21236" target="_blank" rel="noopener noreferrer">PR #21236</a></li>
<li>Anthropic beta headers documentation - <a href="https://github.com/BerriAI/litellm/pull/21320" target="_blank" rel="noopener noreferrer">PR #21320</a></li>
<li>Latency overhead troubleshooting guide - <a href="https://github.com/BerriAI/litellm/pull/21600" target="_blank" rel="noopener noreferrer">PR #21600</a>, <a href="https://github.com/BerriAI/litellm/pull/21603" target="_blank" rel="noopener noreferrer">PR #21603</a></li>
<li>Add rollback safety check guide - <a href="https://github.com/BerriAI/litellm/pull/21743" target="_blank" rel="noopener noreferrer">PR #21743</a></li>
<li>Incident report: vLLM Embeddings broken by encoding_format parameter - <a href="https://github.com/BerriAI/litellm/pull/21474" target="_blank" rel="noopener noreferrer">PR #21474</a></li>
<li>Incident report: Claude Code beta headers - <a href="https://github.com/BerriAI/litellm/pull/21485" target="_blank" rel="noopener noreferrer">PR #21485</a></li>
<li>Mark v1.81.12 as stable - <a href="https://github.com/BerriAI/litellm/pull/21809" target="_blank" rel="noopener noreferrer">PR #21809</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-81-14#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@mjkam made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21306" target="_blank" rel="noopener noreferrer">PR #21306</a></li>
<li>@saneroen made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21243" target="_blank" rel="noopener noreferrer">PR #21243</a></li>
<li>@vincentkoc made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21239" target="_blank" rel="noopener noreferrer">PR #21239</a></li>
<li>@felixti made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19745" target="_blank" rel="noopener noreferrer">PR #19745</a></li>
<li>@anttttti made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20731" target="_blank" rel="noopener noreferrer">PR #20731</a></li>
<li>@ndgigliotti made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21222" target="_blank" rel="noopener noreferrer">PR #21222</a></li>
<li>@iamadamreed made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19912" target="_blank" rel="noopener noreferrer">PR #19912</a></li>
<li>@sahukanishka made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21220" target="_blank" rel="noopener noreferrer">PR #21220</a></li>
<li>@namabile made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21195" target="_blank" rel="noopener noreferrer">PR #21195</a></li>
<li>@stronk7 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21372" target="_blank" rel="noopener noreferrer">PR #21372</a></li>
<li>@ZeroAurora made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21547" target="_blank" rel="noopener noreferrer">PR #21547</a></li>
<li>@SolitudePy made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21497" target="_blank" rel="noopener noreferrer">PR #21497</a></li>
<li>@SherifWaly made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21557" target="_blank" rel="noopener noreferrer">PR #21557</a></li>
<li>@dkindlund made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21633" target="_blank" rel="noopener noreferrer">PR #21633</a></li>
<li>@cagojeiger made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21664" target="_blank" rel="noopener noreferrer">PR #21664</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog">Full Changelog<a href="https://docs.litellm.ai/release_notes/v1-81-14#full-changelog" class="hash-link" aria-label="Direct link to Full Changelog" title="Direct link to Full Changelog">​</a></h2>
<p><a href="https://github.com/BerriAI/litellm/compare/v1.81.12.rc.1...v1.81.14.rc.1" target="_blank" rel="noopener noreferrer">v1.81.12.rc.1...v1.81.14.rc.1</a></p>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[v1.81.12-stable.1 - Guardrail Policy Templates & Action Builder]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-81-12</id>
        <link href="https://docs.litellm.ai/release_notes/v1-81-12"/>
        <updated>2026-02-14T00:00:00.000Z</updated>
        <summary type="html"><![CDATA[Deploy this version]]></summary>
        <content type="html"><![CDATA[<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-81-12#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<!-- -->
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">ghcr.io/berriai/litellm:main-v1.81.12-stable.1</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.81.12</span><br></span></code></pre></div></div></div></div></div>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-81-12#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<ul>
<li><strong>Policy Templates</strong> - <a href="https://docs.litellm.ai/docs/proxy/guardrails/policy_templates">Pre-configured guardrail policy templates for common safety and compliance use-cases (including NSFW, toxic content, and child safety)</a></li>
<li><strong>Guardrail Action Builder</strong> - <a href="https://docs.litellm.ai/docs/proxy/guardrails/policy_templates">Build and customize guardrail policy flows with the new action-builder UI and conditional execution support</a></li>
<li><strong>MCP OAuth2 M2M + Tracing</strong> - <a href="https://docs.litellm.ai/docs/mcp">Add machine-to-machine OAuth2 support for MCP servers and OpenTelemetry tracing for MCP calls through AI Gateway</a></li>
<li><strong>Responses API <code>shell</code> Tool &amp; <code>context_management</code> support</strong> - <a href="https://docs.litellm.ai/docs/response_api">Server-side context management (compaction) and Shell tool support for the OpenAI Responses API</a></li>
<li><strong>Access Groups</strong> - <a href="https://docs.litellm.ai/docs/proxy/access_groups">Create access groups to manage model, MCP server, and agent access across teams and keys</a></li>
<li><strong>50+ New Bedrock Regional Model Entries</strong> - DeepSeek V3.2, MiniMax M2.1, Kimi K2.5, Qwen3 Coder Next, and NVIDIA Nemotron Nano across multiple regions</li>
<li><strong>Add Semgrep &amp; fix OOMs</strong> - <a href="https://docs.litellm.ai/release_notes/v1-81-12#add-semgrep--fix-ooms">Static analysis rules and out-of-memory fixes</a> - <a href="https://github.com/BerriAI/litellm/pull/20912" target="_blank" rel="noopener noreferrer">PR #20912</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="add-semgrep--fix-ooms">Add Semgrep &amp; fix OOMs<a href="https://docs.litellm.ai/release_notes/v1-81-12#add-semgrep--fix-ooms" class="hash-link" aria-label="Direct link to Add Semgrep &amp; fix OOMs" title="Direct link to Add Semgrep &amp; fix OOMs">​</a></h2>
<p>This release fixes out-of-memory (OOM) risks from unbounded <code>asyncio.Queue()</code> usage. Log queues (e.g. GCS bucket) and DB spend-update queues were previously unbounded and could grow without limit under load. They now use a configurable max size (<code>LITELLM_ASYNCIO_QUEUE_MAXSIZE</code>, default 1000); when full, queues flush immediately to make room instead of growing memory. A Semgrep rule (<code>.semgrep/rules/python/unbounded-memory.yml</code>) was added to flag similar unbounded-memory patterns in future code. <a href="https://github.com/BerriAI/litellm/pull/20912" target="_blank" rel="noopener noreferrer">PR #20912</a></p>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="guardrail-action-builder">Guardrail Action Builder<a href="https://docs.litellm.ai/release_notes/v1-81-12#guardrail-action-builder" class="hash-link" aria-label="Direct link to Guardrail Action Builder" title="Direct link to Guardrail Action Builder">​</a></h2>
<p>This release adds a visual action builder for guardrail policies with conditional execution support. You can now chain guardrails into multi-step pipelines — if a simple guardrail fails, route to an advanced one instead of immediately blocking. Each step has configurable ON PASS and ON FAIL actions (Next Step, Block, or Allow), and you can test the full pipeline with a sample message before saving.</p>
<p><img decoding="async" loading="lazy" alt="Guardrail Action Builder" src="https://docs.litellm.ai/assets/images/guard_actions-b7afe0df4cd718760329ede57bfe392f.png" width="2926" height="2000" class="img_ev3q"></p>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="access-groups">Access Groups<a href="https://docs.litellm.ai/release_notes/v1-81-12#access-groups" class="hash-link" aria-label="Direct link to Access Groups" title="Direct link to Access Groups">​</a></h3>
<p>Access Groups simplify defining resource access across your organization. One group can grant access to models, MCP servers, and agents—simply attach it to a key or team. Create groups in the Admin UI, define which resources each group includes, then assign the group when creating keys or teams. Updates to a group apply automatically to all attached keys and teams.</p>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAFCAYAAAB8ZH1oAAAACXBIWXMAACxLAAAsSwGlPZapAAAAfklEQVR4nF2Nuw6CQBQF9///EEJJpUEFZHfvc4xgZTHNyWROmaaJeZ4xM/Z9p9aKueN/lHEcGYaB2iqPZWFbX7gpmfkjgKSIyFkRUZoGRw+6+rm11liPhopRVPW8/eafR3Lbgiqg0hE17s0RiUu8ZKdp8u6JWOJuRCRvdcyCD3N0woPsiF3OAAAAAElFTkSuQmCC&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="334"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/ui_access_groups.8cba803.640.png" srcset="/assets/ideal-img/ui_access_groups.8cba803.640.png 640w,/assets/ideal-img/ui_access_groups.98d4f56.1920.png 1920w" width="640" height="334"></noscript></div>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-providers-and-endpoints">New Providers and Endpoints<a href="https://docs.litellm.ai/release_notes/v1-81-12#new-providers-and-endpoints" class="hash-link" aria-label="Direct link to New Providers and Endpoints" title="Direct link to New Providers and Endpoints">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="new-providers-2-new-providers">New Providers (2 new providers)<a href="https://docs.litellm.ai/release_notes/v1-81-12#new-providers-2-new-providers" class="hash-link" aria-label="Direct link to New Providers (2 new providers)" title="Direct link to New Providers (2 new providers)">​</a></h3>
<table><thead><tr><th>Provider</th><th>Supported LiteLLM Endpoints</th><th>Description</th></tr></thead><tbody><tr><td><a href="https://docs.litellm.ai/docs/providers/scaleway">Scaleway</a></td><td><code>/chat/completions</code></td><td>Scaleway Generative APIs for chat completions</td></tr><tr><td><a href="https://docs.litellm.ai/docs/providers/sarvam">Sarvam AI</a></td><td><code>/chat/completions</code>, <code>/audio/transcriptions</code>, <code>/audio/speech</code></td><td>Sarvam AI STT and TTS support for Indian languages</td></tr></tbody></table>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-81-12#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support-19-highlighted-models">New Model Support (19 highlighted models)<a href="https://docs.litellm.ai/release_notes/v1-81-12#new-model-support-19-highlighted-models" class="hash-link" aria-label="Direct link to New Model Support (19 highlighted models)" title="Direct link to New Model Support (19 highlighted models)">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th></tr></thead><tbody><tr><td>AWS Bedrock</td><td><code>deepseek.v3.2</code></td><td>164K</td><td>$0.62</td><td>$1.85</td></tr><tr><td>AWS Bedrock</td><td><code>minimax.minimax-m2.1</code></td><td>196K</td><td>$0.30</td><td>$1.20</td></tr><tr><td>AWS Bedrock</td><td><code>moonshotai.kimi-k2.5</code></td><td>262K</td><td>$0.60</td><td>$3.00</td></tr><tr><td>AWS Bedrock</td><td><code>moonshotai.kimi-k2-thinking</code></td><td>262K</td><td>$0.73</td><td>$3.03</td></tr><tr><td>AWS Bedrock</td><td><code>qwen.qwen3-coder-next</code></td><td>262K</td><td>$0.50</td><td>$1.20</td></tr><tr><td>AWS Bedrock</td><td><code>nvidia.nemotron-nano-3-30b</code></td><td>262K</td><td>$0.06</td><td>$0.24</td></tr><tr><td>Azure AI</td><td><code>azure_ai/kimi-k2.5</code></td><td>262K</td><td>$0.60</td><td>$3.00</td></tr><tr><td>Vertex AI</td><td><code>vertex_ai/zai-org/glm-5-maas</code></td><td>200K</td><td>$1.00</td><td>$3.20</td></tr><tr><td>MiniMax</td><td><code>minimax/MiniMax-M2.5</code></td><td>1M</td><td>$0.30</td><td>$1.20</td></tr><tr><td>MiniMax</td><td><code>minimax/MiniMax-M2.5-lightning</code></td><td>1M</td><td>$0.30</td><td>$2.40</td></tr><tr><td>Dashscope</td><td><code>dashscope/qwen3-max</code></td><td>258K</td><td>Tiered pricing</td><td>Tiered pricing</td></tr><tr><td>Perplexity</td><td><code>perplexity/preset/pro-search</code></td><td>-</td><td>Per-request</td><td>Per-request</td></tr><tr><td>Perplexity</td><td><code>perplexity/openai/gpt-4o</code></td><td>-</td><td>Per-request</td><td>Per-request</td></tr><tr><td>Perplexity</td><td><code>perplexity/openai/gpt-5.2</code></td><td>-</td><td>Per-request</td><td>Per-request</td></tr><tr><td>Vercel AI Gateway</td><td><code>vercel_ai_gateway/anthropic/claude-opus-4.6</code></td><td>200K</td><td>$5.00</td><td>$25.00</td></tr><tr><td>Vercel AI Gateway</td><td><code>vercel_ai_gateway/anthropic/claude-sonnet-4</code></td><td>200K</td><td>$3.00</td><td>$15.00</td></tr><tr><td>Vercel AI Gateway</td><td><code>vercel_ai_gateway/anthropic/claude-haiku-4.5</code></td><td>200K</td><td>$1.00</td><td>$5.00</td></tr><tr><td>Sarvam AI</td><td><code>sarvam/sarvam-m</code></td><td>8K</td><td>Free tier</td><td>Free tier</td></tr><tr><td>Anthropic</td><td><code>fast/claude-opus-4-6</code></td><td>1M</td><td>$30.00</td><td>$150.00</td></tr></tbody></table>
<p><em>Note: AWS Bedrock models are available across multiple regions (us-east-1, us-east-2, us-west-2, eu-central-1, eu-north-1, ap-northeast-1, ap-south-1, ap-southeast-3, sa-east-1). 54 regional model entries were added in total.</em></p>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-81-12#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Enable non-tool structured outputs on Claude Opus 4.5 and 4.6 using <code>output_format</code> param - <a href="https://github.com/BerriAI/litellm/pull/20548" target="_blank" rel="noopener noreferrer">PR #20548</a></li>
<li>Add support for <code>anthropic_messages</code> call type in prompt caching - <a href="https://github.com/BerriAI/litellm/pull/19233" target="_blank" rel="noopener noreferrer">PR #19233</a></li>
<li>Managing Anthropic Beta Headers with remote URL fetching - <a href="https://github.com/BerriAI/litellm/pull/20935" target="_blank" rel="noopener noreferrer">PR #20935</a>, <a href="https://github.com/BerriAI/litellm/pull/21110" target="_blank" rel="noopener noreferrer">PR #21110</a></li>
<li>Remove <code>x-anthropic-billing</code> block - <a href="https://github.com/BerriAI/litellm/pull/20951" target="_blank" rel="noopener noreferrer">PR #20951</a></li>
<li>Use Authorization Bearer for OAuth tokens instead of x-api-key - <a href="https://github.com/BerriAI/litellm/pull/21039" target="_blank" rel="noopener noreferrer">PR #21039</a></li>
<li>Filter unsupported JSON schema constraints for structured outputs - <a href="https://github.com/BerriAI/litellm/pull/20813" target="_blank" rel="noopener noreferrer">PR #20813</a></li>
<li>New Claude Opus 4.6 features for <code>/v1/messages</code> - <a href="https://github.com/BerriAI/litellm/pull/20733" target="_blank" rel="noopener noreferrer">PR #20733</a></li>
<li>Fix <code>reasoning_effort=None</code> and <code>"none"</code> should return None for Opus 4.6 - <a href="https://github.com/BerriAI/litellm/pull/20800" target="_blank" rel="noopener noreferrer">PR #20800</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">AWS Bedrock</a></strong></p>
<ul>
<li>Extend model support with 4 new beta models - <a href="https://github.com/BerriAI/litellm/pull/21035" target="_blank" rel="noopener noreferrer">PR #21035</a></li>
<li>Add Claude Opus 4.6 to <code>_supports_tool_search_on_bedrock</code> - <a href="https://github.com/BerriAI/litellm/pull/21017" target="_blank" rel="noopener noreferrer">PR #21017</a></li>
<li>Correct Bedrock Claude Opus 4.6 model IDs (remove <code>:0</code> suffix) - <a href="https://github.com/BerriAI/litellm/pull/20564" target="_blank" rel="noopener noreferrer">PR #20564</a>, <a href="https://github.com/BerriAI/litellm/pull/20671" target="_blank" rel="noopener noreferrer">PR #20671</a></li>
<li>Add <code>output_config</code> as supported param - <a href="https://github.com/BerriAI/litellm/pull/20748" target="_blank" rel="noopener noreferrer">PR #20748</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vertex">Vertex AI</a></strong></p>
<ul>
<li>Add Vertex GLM-5 model support - <a href="https://github.com/BerriAI/litellm/pull/21053" target="_blank" rel="noopener noreferrer">PR #21053</a></li>
<li>Propagate <code>extra_headers</code> anthropic-beta to request body - <a href="https://github.com/BerriAI/litellm/pull/20666" target="_blank" rel="noopener noreferrer">PR #20666</a></li>
<li>Preserve <code>usageMetadata</code> in <code>_hidden_params</code> - <a href="https://github.com/BerriAI/litellm/pull/20559" target="_blank" rel="noopener noreferrer">PR #20559</a></li>
<li>Map <code>IMAGE_PROHIBITED_CONTENT</code> to <code>content_filter</code> - <a href="https://github.com/BerriAI/litellm/pull/20524" target="_blank" rel="noopener noreferrer">PR #20524</a></li>
<li>Add RAG ingest for Vertex AI - <a href="https://github.com/BerriAI/litellm/pull/21120" target="_blank" rel="noopener noreferrer">PR #21120</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/cohere">OCI / Cohere</a></strong></p>
<ul>
<li>OCI Cohere responseFormat/Pydantic support - <a href="https://github.com/BerriAI/litellm/pull/20663" target="_blank" rel="noopener noreferrer">PR #20663</a></li>
<li>Fix OCI Cohere system messages by populating <code>preambleOverride</code> - <a href="https://github.com/BerriAI/litellm/pull/20958" target="_blank" rel="noopener noreferrer">PR #20958</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/perplexity">Perplexity</a></strong></p>
<ul>
<li>Perplexity Research API support with preset search - <a href="https://github.com/BerriAI/litellm/pull/20860" target="_blank" rel="noopener noreferrer">PR #20860</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/minimax">MiniMax</a></strong></p>
<ul>
<li>Add MiniMax-M2.5 and MiniMax-M2.5-lightning models - <a href="https://github.com/BerriAI/litellm/pull/21054" target="_blank" rel="noopener noreferrer">PR #21054</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/moonshot">Kimi / Moonshot</a></strong></p>
<ul>
<li>Add Kimi model pricing by region - <a href="https://github.com/BerriAI/litellm/pull/20855" target="_blank" rel="noopener noreferrer">PR #20855</a></li>
<li>Add <code>moonshotai.kimi-k2.5</code> - <a href="https://github.com/BerriAI/litellm/pull/20863" target="_blank" rel="noopener noreferrer">PR #20863</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/dashscope">Dashscope</a></strong></p>
<ul>
<li>Add <code>dashscope/qwen3-max</code> model with tiered pricing - <a href="https://github.com/BerriAI/litellm/pull/20919" target="_blank" rel="noopener noreferrer">PR #20919</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vercel_ai_gateway">Vercel AI Gateway</a></strong></p>
<ul>
<li>Add new Vercel AI Anthropic models - <a href="https://github.com/BerriAI/litellm/pull/20745" target="_blank" rel="noopener noreferrer">PR #20745</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/azure_ai">Azure AI</a></strong></p>
<ul>
<li>Add <code>azure_ai/kimi-k2.5</code> to Azure model DB - <a href="https://github.com/BerriAI/litellm/pull/20896" target="_blank" rel="noopener noreferrer">PR #20896</a></li>
<li>Support Azure AD token auth for non-Claude azure_ai models - <a href="https://github.com/BerriAI/litellm/pull/20981" target="_blank" rel="noopener noreferrer">PR #20981</a></li>
<li>Fix Azure batches issues - <a href="https://github.com/BerriAI/litellm/pull/21092" target="_blank" rel="noopener noreferrer">PR #21092</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/deepseek">DeepSeek</a></strong></p>
<ul>
<li>Sync DeepSeek model metadata and add bare-name fallback - <a href="https://github.com/BerriAI/litellm/pull/20938" target="_blank" rel="noopener noreferrer">PR #20938</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Gemini</a></strong></p>
<ul>
<li>Handle image in assistant message for Gemini - <a href="https://github.com/BerriAI/litellm/pull/20845" target="_blank" rel="noopener noreferrer">PR #20845</a></li>
<li>Add missing tpm/rpm for Gemini models - <a href="https://github.com/BerriAI/litellm/pull/21175" target="_blank" rel="noopener noreferrer">PR #21175</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Add 30 missing models to pricing JSON - <a href="https://github.com/BerriAI/litellm/pull/20797" target="_blank" rel="noopener noreferrer">PR #20797</a></li>
<li>Cleanup 39 deprecated OpenRouter models - <a href="https://github.com/BerriAI/litellm/pull/20786" target="_blank" rel="noopener noreferrer">PR #20786</a></li>
<li>Standardize endpoint <code>display_name</code> naming convention - <a href="https://github.com/BerriAI/litellm/pull/20791" target="_blank" rel="noopener noreferrer">PR #20791</a></li>
<li>Fix and stabilize model cost map formatting - <a href="https://github.com/BerriAI/litellm/pull/20895" target="_blank" rel="noopener noreferrer">PR #20895</a></li>
<li>Export <code>PermissionDeniedError</code> from <code>litellm.__init__</code> - <a href="https://github.com/BerriAI/litellm/pull/20960" target="_blank" rel="noopener noreferrer">PR #20960</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-81-12#bug-fixes" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Fix <code>get_supported_anthropic_messages_params</code> - <a href="https://github.com/BerriAI/litellm/pull/20752" target="_blank" rel="noopener noreferrer">PR #20752</a></li>
<li>Fix <code>base_model</code> name for body and deployment name in URL - <a href="https://github.com/BerriAI/litellm/pull/20747" target="_blank" rel="noopener noreferrer">PR #20747</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/azure/azure">Azure</a></strong></p>
<ul>
<li>Preserve <code>content_policy_violation</code> error details from Azure OpenAI - <a href="https://github.com/BerriAI/litellm/pull/20883" target="_blank" rel="noopener noreferrer">PR #20883</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vertex">Vertex AI</a></strong></p>
<ul>
<li>Fix Gemini multi-turn tool calling message formatting (added and reverted) - <a href="https://github.com/BerriAI/litellm/pull/20569" target="_blank" rel="noopener noreferrer">PR #20569</a>, <a href="https://github.com/BerriAI/litellm/pull/21051" target="_blank" rel="noopener noreferrer">PR #21051</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-81-12#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-81-12#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/response_api">Responses API</a></strong></p>
<ul>
<li>Add server-side context management (compaction) support - <a href="https://github.com/BerriAI/litellm/pull/21058" target="_blank" rel="noopener noreferrer">PR #21058</a></li>
<li>Add Shell tool support for OpenAI Responses API - <a href="https://github.com/BerriAI/litellm/pull/21063" target="_blank" rel="noopener noreferrer">PR #21063</a></li>
<li>Preserve tool call argument deltas when streaming id is omitted - <a href="https://github.com/BerriAI/litellm/pull/20712" target="_blank" rel="noopener noreferrer">PR #20712</a></li>
<li>Preserve interleaved thinking/redacted_thinking blocks during streaming - <a href="https://github.com/BerriAI/litellm/pull/20702" target="_blank" rel="noopener noreferrer">PR #20702</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/completion/input">Chat Completions</a></strong></p>
<ul>
<li>Add Web Search support using LiteLLM <code>/search</code> (web search interception hook) - <a href="https://github.com/BerriAI/litellm/pull/20483" target="_blank" rel="noopener noreferrer">PR #20483</a></li>
<li>Preserved nullable object fields by carrying schema properties - <a href="https://github.com/BerriAI/litellm/pull/19132" target="_blank" rel="noopener noreferrer">PR #19132</a></li>
<li>Support <code>prompt_cache_key</code> for OpenAI and Azure chat completions - <a href="https://github.com/BerriAI/litellm/pull/20989" target="_blank" rel="noopener noreferrer">PR #20989</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/pass_through/bedrock">Pass-Through Endpoints</a></strong></p>
<ul>
<li>Add support for <code>langchain_aws</code> via LiteLLM passthrough - <a href="https://github.com/BerriAI/litellm/pull/20843" target="_blank" rel="noopener noreferrer">PR #20843</a></li>
<li>Add <code>custom_body</code> parameter to <code>endpoint_func</code> in <code>create_pass_through_route</code> - <a href="https://github.com/BerriAI/litellm/pull/20849" target="_blank" rel="noopener noreferrer">PR #20849</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openai">Vector Stores</a></strong></p>
<ul>
<li>Add <code>target_model_names</code> for vector store endpoints - <a href="https://github.com/BerriAI/litellm/pull/21089" target="_blank" rel="noopener noreferrer">PR #21089</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Add <code>output_config</code> as supported param - <a href="https://github.com/BerriAI/litellm/pull/20748" target="_blank" rel="noopener noreferrer">PR #20748</a></li>
<li>Add managed error file support - <a href="https://github.com/BerriAI/litellm/pull/20838" target="_blank" rel="noopener noreferrer">PR #20838</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs">Bugs<a href="https://docs.litellm.ai/release_notes/v1-81-12#bugs" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>General</strong>
<ul>
<li>Stop leaking Python tracebacks in streaming SSE error responses - <a href="https://github.com/BerriAI/litellm/pull/20850" target="_blank" rel="noopener noreferrer">PR #20850</a></li>
<li>Fix video list pagination cursors not encoded with provider metadata - <a href="https://github.com/BerriAI/litellm/pull/20710" target="_blank" rel="noopener noreferrer">PR #20710</a></li>
<li>Handle <code>metadata=None</code> in SDK path retry/error logic - <a href="https://github.com/BerriAI/litellm/pull/20873" target="_blank" rel="noopener noreferrer">PR #20873</a></li>
<li>Fix Spend logs pickle error with Pydantic models and redaction - <a href="https://github.com/BerriAI/litellm/pull/20685" target="_blank" rel="noopener noreferrer">PR #20685</a></li>
<li>Remove duplicate <code>PerplexityResponsesConfig</code> from <code>LLM_CONFIG_NAMES</code> - <a href="https://github.com/BerriAI/litellm/pull/21105" target="_blank" rel="noopener noreferrer">PR #21105</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-81-12#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-81-12#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong>Access Groups</strong></p>
<ul>
<li>New Access Groups feature for managing model, MCP server, and agent access - <a href="https://github.com/BerriAI/litellm/pull/21022" target="_blank" rel="noopener noreferrer">PR #21022</a></li>
<li>Access Groups table and details page UI - <a href="https://github.com/BerriAI/litellm/pull/21165" target="_blank" rel="noopener noreferrer">PR #21165</a></li>
<li>Refactor <code>model_ids</code> to <code>model_names</code> for backwards compatibility - <a href="https://github.com/BerriAI/litellm/pull/21166" target="_blank" rel="noopener noreferrer">PR #21166</a></li>
</ul>
</li>
<li>
<p><strong>Policies</strong></p>
<ul>
<li>Allow connecting Policies to Tags, simulating Policies, viewing key/team counts - <a href="https://github.com/BerriAI/litellm/pull/20904" target="_blank" rel="noopener noreferrer">PR #20904</a></li>
<li>Guardrail pipeline support for conditional sequential execution - <a href="https://github.com/BerriAI/litellm/pull/21177" target="_blank" rel="noopener noreferrer">PR #21177</a></li>
<li>Pipeline flow builder UI for guardrail policies - <a href="https://github.com/BerriAI/litellm/pull/21188" target="_blank" rel="noopener noreferrer">PR #21188</a></li>
</ul>
</li>
<li>
<p><strong>SSO / Auth</strong></p>
<ul>
<li>New Login With SSO Button - <a href="https://github.com/BerriAI/litellm/pull/20908" target="_blank" rel="noopener noreferrer">PR #20908</a></li>
<li>M2M OAuth2 UI Flow - <a href="https://github.com/BerriAI/litellm/pull/20794" target="_blank" rel="noopener noreferrer">PR #20794</a></li>
<li>Allow Organization and Team Admins to call <code>/invitation/new</code> - <a href="https://github.com/BerriAI/litellm/pull/20987" target="_blank" rel="noopener noreferrer">PR #20987</a></li>
<li>Invite User: Email Integration Alert - <a href="https://github.com/BerriAI/litellm/pull/20790" target="_blank" rel="noopener noreferrer">PR #20790</a></li>
<li>Populate identity fields in proxy admin JWT early-return path - <a href="https://github.com/BerriAI/litellm/pull/21169" target="_blank" rel="noopener noreferrer">PR #21169</a></li>
</ul>
</li>
<li>
<p><strong>Spend Logs</strong></p>
<ul>
<li>Show predefined error codes in filter with user definable fallback - <a href="https://github.com/BerriAI/litellm/pull/20773" target="_blank" rel="noopener noreferrer">PR #20773</a></li>
<li>Paginated searchable model select - <a href="https://github.com/BerriAI/litellm/pull/20892" target="_blank" rel="noopener noreferrer">PR #20892</a></li>
<li>Sorting columns support - <a href="https://github.com/BerriAI/litellm/pull/21143" target="_blank" rel="noopener noreferrer">PR #21143</a></li>
<li>Allow sorting on <code>/spend/logs/ui</code> - <a href="https://github.com/BerriAI/litellm/pull/20991" target="_blank" rel="noopener noreferrer">PR #20991</a></li>
</ul>
</li>
<li>
<p><strong>UI Improvements</strong></p>
<ul>
<li>Navbar: Option to hide Usage Popup - <a href="https://github.com/BerriAI/litellm/pull/20910" target="_blank" rel="noopener noreferrer">PR #20910</a></li>
<li>Model Page: Improve Credentials Messaging - <a href="https://github.com/BerriAI/litellm/pull/21076" target="_blank" rel="noopener noreferrer">PR #21076</a></li>
<li>Fallbacks: Default configurable to 10 models - <a href="https://github.com/BerriAI/litellm/pull/21144" target="_blank" rel="noopener noreferrer">PR #21144</a></li>
<li>Fallback display with arrows and card structure - <a href="https://github.com/BerriAI/litellm/pull/20922" target="_blank" rel="noopener noreferrer">PR #20922</a></li>
<li>Team Info: Migrate to AntD Tabs + Table - <a href="https://github.com/BerriAI/litellm/pull/20785" target="_blank" rel="noopener noreferrer">PR #20785</a></li>
<li>AntD refactoring and 0 cost models fix - <a href="https://github.com/BerriAI/litellm/pull/20687" target="_blank" rel="noopener noreferrer">PR #20687</a></li>
<li>Zscaler AI Guard UI - <a href="https://github.com/BerriAI/litellm/pull/21077" target="_blank" rel="noopener noreferrer">PR #21077</a></li>
<li>Include Config Defined Pass Through Endpoints - <a href="https://github.com/BerriAI/litellm/pull/20898" target="_blank" rel="noopener noreferrer">PR #20898</a></li>
<li>Rename "HTTP" to "Streamable HTTP (Recommended)" in MCP server page - <a href="https://github.com/BerriAI/litellm/pull/21000" target="_blank" rel="noopener noreferrer">PR #21000</a></li>
<li>MCP server discovery UI - <a href="https://github.com/BerriAI/litellm/pull/21079" target="_blank" rel="noopener noreferrer">PR #21079</a></li>
</ul>
</li>
<li>
<p><strong>Virtual Keys</strong></p>
<ul>
<li>Allow Management keys to access <code>user/daily/activity</code> and team - <a href="https://github.com/BerriAI/litellm/pull/20124" target="_blank" rel="noopener noreferrer">PR #20124</a></li>
<li>Skip premium check for empty metadata fields on team/key update - <a href="https://github.com/BerriAI/litellm/pull/20598" target="_blank" rel="noopener noreferrer">PR #20598</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs-1">Bugs<a href="https://docs.litellm.ai/release_notes/v1-81-12#bugs-1" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li>Logs: Fix Input and Output Copying - <a href="https://github.com/BerriAI/litellm/pull/20657" target="_blank" rel="noopener noreferrer">PR #20657</a></li>
<li>Teams: Fix Available Teams - <a href="https://github.com/BerriAI/litellm/pull/20682" target="_blank" rel="noopener noreferrer">PR #20682</a></li>
<li>Spend Logs: Reset Filters Resets Custom Date Range - <a href="https://github.com/BerriAI/litellm/pull/21149" target="_blank" rel="noopener noreferrer">PR #21149</a></li>
<li>Usage: Request Chart stack variant fix - <a href="https://github.com/BerriAI/litellm/pull/20894" target="_blank" rel="noopener noreferrer">PR #20894</a></li>
<li>Add Auto Router: Description Text Input Focus - <a href="https://github.com/BerriAI/litellm/pull/21004" target="_blank" rel="noopener noreferrer">PR #21004</a></li>
<li>Guardrail Edit: LiteLLM Content Filter Categories - <a href="https://github.com/BerriAI/litellm/pull/21002" target="_blank" rel="noopener noreferrer">PR #21002</a></li>
<li>Add null guard for models in API keys table - <a href="https://github.com/BerriAI/litellm/pull/20655" target="_blank" rel="noopener noreferrer">PR #20655</a></li>
<li>Show error details instead of 'Data Not Available' for failed requests - <a href="https://github.com/BerriAI/litellm/pull/20656" target="_blank" rel="noopener noreferrer">PR #20656</a></li>
<li>Fix Spend Management Tests - <a href="https://github.com/BerriAI/litellm/pull/21088" target="_blank" rel="noopener noreferrer">PR #21088</a></li>
<li>Fix JWT email domain validation error message - <a href="https://github.com/BerriAI/litellm/pull/21212" target="_blank" rel="noopener noreferrer">PR #21212</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="ai-integrations">AI Integrations<a href="https://docs.litellm.ai/release_notes/v1-81-12#ai-integrations" class="hash-link" aria-label="Direct link to AI Integrations" title="Direct link to AI Integrations">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="logging">Logging<a href="https://docs.litellm.ai/release_notes/v1-81-12#logging" class="hash-link" aria-label="Direct link to Logging" title="Direct link to Logging">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/observability/posthog_integration">PostHog</a></strong></p>
<ul>
<li>Fix JSON serialization error for non-serializable objects - <a href="https://github.com/BerriAI/litellm/pull/20668" target="_blank" rel="noopener noreferrer">PR #20668</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#prometheus">Prometheus</a></strong></p>
<ul>
<li>Sanitize label values to prevent metric scrape failures - <a href="https://github.com/BerriAI/litellm/pull/20600" target="_blank" rel="noopener noreferrer">PR #20600</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#langfuse">Langfuse</a></strong></p>
<ul>
<li>Prevent empty proxy request spans from being sent to Langfuse - <a href="https://github.com/BerriAI/litellm/pull/19935" target="_blank" rel="noopener noreferrer">PR #19935</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#otel">OpenTelemetry</a></strong></p>
<ul>
<li>Auto-infer <code>otlp_http</code> exporter when endpoint is configured - <a href="https://github.com/BerriAI/litellm/pull/20438" target="_blank" rel="noopener noreferrer">PR #20438</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging">CloudZero</a></strong></p>
<ul>
<li>Update CBF field mappings per LIT-1907 - <a href="https://github.com/BerriAI/litellm/pull/20906" target="_blank" rel="noopener noreferrer">PR #20906</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Allow <code>MAX_CALLBACKS</code> override via env var - <a href="https://github.com/BerriAI/litellm/pull/20781" target="_blank" rel="noopener noreferrer">PR #20781</a></li>
<li>Add <code>standard_logging_payload_excluded_fields</code> config option - <a href="https://github.com/BerriAI/litellm/pull/20831" target="_blank" rel="noopener noreferrer">PR #20831</a></li>
<li>Enable <code>verbose_logger</code> when <code>LITELLM_LOG=DEBUG</code> - <a href="https://github.com/BerriAI/litellm/pull/20496" target="_blank" rel="noopener noreferrer">PR #20496</a></li>
<li>Guard against None <code>litellm_metadata</code> in batch logging path - <a href="https://github.com/BerriAI/litellm/pull/20832" target="_blank" rel="noopener noreferrer">PR #20832</a></li>
<li>Propagate model-level tags from config to SpendLogs - <a href="https://github.com/BerriAI/litellm/pull/20769" target="_blank" rel="noopener noreferrer">PR #20769</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails">Guardrails<a href="https://docs.litellm.ai/release_notes/v1-81-12#guardrails" class="hash-link" aria-label="Direct link to Guardrails" title="Direct link to Guardrails">​</a></h3>
<ul>
<li>
<p><strong>Policy Templates</strong></p>
<ul>
<li>New Policy Templates: pre-configured guardrail combinations for specific use-cases - <a href="https://github.com/BerriAI/litellm/pull/21025" target="_blank" rel="noopener noreferrer">PR #21025</a></li>
<li>Add NSFW policy template, toxic keywords in multiple languages, child safety content filter, JSON content viewer - <a href="https://github.com/BerriAI/litellm/pull/21205" target="_blank" rel="noopener noreferrer">PR #21205</a></li>
<li>Add toxic/abusive content filter guardrails - <a href="https://github.com/BerriAI/litellm/pull/20934" target="_blank" rel="noopener noreferrer">PR #20934</a></li>
</ul>
</li>
<li>
<p><strong>Pipeline Execution</strong></p>
<ul>
<li>Add guardrail pipeline support for conditional sequential execution - <a href="https://github.com/BerriAI/litellm/pull/21177" target="_blank" rel="noopener noreferrer">PR #21177</a></li>
<li>Agent Guardrails on streaming output - <a href="https://github.com/BerriAI/litellm/pull/21206" target="_blank" rel="noopener noreferrer">PR #21206</a></li>
<li>Pipeline flow builder UI - <a href="https://github.com/BerriAI/litellm/pull/21188" target="_blank" rel="noopener noreferrer">PR #21188</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/apply_guardrail">Zscaler AI Guard</a></strong></p>
<ul>
<li>Zscaler AI Guard bug fixes and support during post-call - <a href="https://github.com/BerriAI/litellm/pull/20801" target="_blank" rel="noopener noreferrer">PR #20801</a></li>
<li>Zscaler AI Guard UI - <a href="https://github.com/BerriAI/litellm/pull/21077" target="_blank" rel="noopener noreferrer">PR #21077</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/apply_guardrail">ZGuard</a></strong></p>
<ul>
<li>Add team policy mapping for ZGuard - <a href="https://github.com/BerriAI/litellm/pull/20608" target="_blank" rel="noopener noreferrer">PR #20608</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Add logging to all unified guardrails + link to custom code guardrail templates - <a href="https://github.com/BerriAI/litellm/pull/20900" target="_blank" rel="noopener noreferrer">PR #20900</a></li>
<li>Forward request headers + <code>litellm_version</code> to generic guardrails - <a href="https://github.com/BerriAI/litellm/pull/20729" target="_blank" rel="noopener noreferrer">PR #20729</a></li>
<li>Empty <code>guardrails</code>/<code>policies</code> arrays should not trigger enterprise license check - <a href="https://github.com/BerriAI/litellm/pull/20567" target="_blank" rel="noopener noreferrer">PR #20567</a></li>
<li>Fix OpenAI moderation guardrails - <a href="https://github.com/BerriAI/litellm/pull/20718" target="_blank" rel="noopener noreferrer">PR #20718</a></li>
<li>Fix <code>/v2/guardrails/list</code> returning sensitive values - <a href="https://github.com/BerriAI/litellm/pull/20796" target="_blank" rel="noopener noreferrer">PR #20796</a></li>
<li>Fix guardrail status error - <a href="https://github.com/BerriAI/litellm/pull/20972" target="_blank" rel="noopener noreferrer">PR #20972</a></li>
<li>Reuse <code>get_instance_fn</code> in <code>initialize_custom_guardrail</code> - <a href="https://github.com/BerriAI/litellm/pull/20917" target="_blank" rel="noopener noreferrer">PR #20917</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-81-12#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li><strong>Prevent shared backend model key from being polluted</strong> by per-deployment custom pricing - <a href="https://github.com/BerriAI/litellm/pull/20679" target="_blank" rel="noopener noreferrer">PR #20679</a></li>
<li><strong>Avoid in-place mutation</strong> in SpendUpdateQueue aggregation - <a href="https://github.com/BerriAI/litellm/pull/20876" target="_blank" rel="noopener noreferrer">PR #20876</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-gateway-12-updates">MCP Gateway (12 updates)<a href="https://docs.litellm.ai/release_notes/v1-81-12#mcp-gateway-12-updates" class="hash-link" aria-label="Direct link to MCP Gateway (12 updates)" title="Direct link to MCP Gateway (12 updates)">​</a></h2>
<ul>
<li><strong>MCP M2M OAuth2 Support</strong> - Add support for machine-to-machine OAuth2 for MCP servers - <a href="https://github.com/BerriAI/litellm/pull/20788" target="_blank" rel="noopener noreferrer">PR #20788</a></li>
<li><strong>MCP Server Discovery UI</strong> - Browse and discover available MCP servers from the UI - <a href="https://github.com/BerriAI/litellm/pull/21079" target="_blank" rel="noopener noreferrer">PR #21079</a></li>
<li><strong>MCP Tracing</strong> - Add OpenTelemetry tracing for MCP calls running through AI Gateway - <a href="https://github.com/BerriAI/litellm/pull/21018" target="_blank" rel="noopener noreferrer">PR #21018</a></li>
<li><strong>MCP OAuth2 Debug Headers</strong> - Client-side debug headers for OAuth2 troubleshooting - <a href="https://github.com/BerriAI/litellm/pull/21151" target="_blank" rel="noopener noreferrer">PR #21151</a></li>
<li><strong>Fix MCP "Session not found" errors</strong> - Resolve session persistence issues - <a href="https://github.com/BerriAI/litellm/pull/21040" target="_blank" rel="noopener noreferrer">PR #21040</a></li>
<li><strong>Fix MCP OAuth2 root endpoints</strong> returning "MCP server not found" - <a href="https://github.com/BerriAI/litellm/pull/20784" target="_blank" rel="noopener noreferrer">PR #20784</a></li>
<li><strong>Fix MCP OAuth2 query param merging</strong> when <code>authorization_url</code> already contains params - <a href="https://github.com/BerriAI/litellm/pull/20968" target="_blank" rel="noopener noreferrer">PR #20968</a></li>
<li><strong>Fix MCP SCOPES on Atlassian</strong> issue - <a href="https://github.com/BerriAI/litellm/pull/21150" target="_blank" rel="noopener noreferrer">PR #21150</a></li>
<li><strong>Fix MCP StreamableHTTP backend</strong> - Use <code>anyio.fail_after</code> instead of <code>asyncio.wait_for</code> - <a href="https://github.com/BerriAI/litellm/pull/20891" target="_blank" rel="noopener noreferrer">PR #20891</a></li>
<li><strong>Inject <code>NPM_CONFIG_CACHE</code></strong> into STDIO MCP subprocess env - <a href="https://github.com/BerriAI/litellm/pull/21069" target="_blank" rel="noopener noreferrer">PR #21069</a></li>
<li><strong>Block spaces and hyphens</strong> in MCP server names and aliases - <a href="https://github.com/BerriAI/litellm/pull/21074" target="_blank" rel="noopener noreferrer">PR #21074</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements-8-improvements">Performance / Loadbalancing / Reliability improvements (8 improvements)<a href="https://docs.litellm.ai/release_notes/v1-81-12#performance--loadbalancing--reliability-improvements-8-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements (8 improvements)" title="Direct link to Performance / Loadbalancing / Reliability improvements (8 improvements)">​</a></h2>
<ul>
<li><strong>Remove orphan entries from queue</strong> - Fix memory leak in scheduler queue - <a href="https://github.com/BerriAI/litellm/pull/20866" target="_blank" rel="noopener noreferrer">PR #20866</a></li>
<li><strong>Remove repeated provider parsing</strong> in budget limiter hot path - <a href="https://github.com/BerriAI/litellm/pull/21043" target="_blank" rel="noopener noreferrer">PR #21043</a></li>
<li><strong>Use current retry exception</strong> for retry backoff instead of stale exception - <a href="https://github.com/BerriAI/litellm/pull/20725" target="_blank" rel="noopener noreferrer">PR #20725</a></li>
<li><strong>Add Semgrep &amp; fix OOMs</strong> - Static analysis rules and out-of-memory fixes - <a href="https://github.com/BerriAI/litellm/pull/20912" target="_blank" rel="noopener noreferrer">PR #20912</a></li>
<li><strong>Add Pyroscope</strong> for continuous profiling and observability - <a href="https://github.com/BerriAI/litellm/pull/21167" target="_blank" rel="noopener noreferrer">PR #21167</a></li>
<li><strong>Respect <code>ssl_verify</code></strong> with shared aiohttp sessions - <a href="https://github.com/BerriAI/litellm/pull/20349" target="_blank" rel="noopener noreferrer">PR #20349</a></li>
<li><strong>Fix shared health check serialization</strong> - <a href="https://github.com/BerriAI/litellm/pull/21119" target="_blank" rel="noopener noreferrer">PR #21119</a></li>
<li><strong>Change model mismatch logs</strong> from WARNING to DEBUG - <a href="https://github.com/BerriAI/litellm/pull/20994" target="_blank" rel="noopener noreferrer">PR #20994</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="database-changes">Database Changes<a href="https://docs.litellm.ai/release_notes/v1-81-12#database-changes" class="hash-link" aria-label="Direct link to Database Changes" title="Direct link to Database Changes">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="schema-updates">Schema Updates<a href="https://docs.litellm.ai/release_notes/v1-81-12#schema-updates" class="hash-link" aria-label="Direct link to Schema Updates" title="Direct link to Schema Updates">​</a></h3>
<table><thead><tr><th>Table</th><th>Change Type</th><th>Description</th><th>PR</th><th>Migration</th></tr></thead><tbody><tr><td><code>LiteLLM_VerificationToken</code></td><td>New Indexes</td><td>Added indexes on <code>user_id</code>+<code>team_id</code>, <code>team_id</code>, and <code>budget_reset_at</code>+<code>expires</code></td><td><a href="https://github.com/BerriAI/litellm/pull/20736" target="_blank" rel="noopener noreferrer">PR #20736</a></td><td><a href="https://github.com/BerriAI/litellm/blob/main/litellm-proxy-extras/litellm_proxy_extras/migrations/20260209085821_add_verificationtoken_indexes/migration.sql" target="_blank" rel="noopener noreferrer">Migration</a></td></tr><tr><td><code>LiteLLM_PolicyAttachmentTable</code></td><td>New Column</td><td>Added <code>tags</code> text array for policy-to-tag connections</td><td><a href="https://github.com/BerriAI/litellm/pull/21061" target="_blank" rel="noopener noreferrer">PR #21061</a></td><td><a href="https://github.com/BerriAI/litellm/blob/main/litellm-proxy-extras/litellm_proxy_extras/migrations/20260212103349_adjust_tags_policy_table/migration.sql" target="_blank" rel="noopener noreferrer">Migration</a></td></tr><tr><td><code>LiteLLM_AccessGroupTable</code></td><td>New Table</td><td>Access groups for managing model, MCP server, and agent access</td><td><a href="https://github.com/BerriAI/litellm/pull/21022" target="_blank" rel="noopener noreferrer">PR #21022</a></td><td><a href="https://github.com/BerriAI/litellm/blob/main/litellm-proxy-extras/litellm_proxy_extras/migrations/20260212143306_add_access_group_table/migration.sql" target="_blank" rel="noopener noreferrer">Migration</a></td></tr><tr><td><code>LiteLLM_AccessGroupTable</code></td><td>Column Change</td><td>Renamed <code>access_model_ids</code> to <code>access_model_names</code></td><td><a href="https://github.com/BerriAI/litellm/pull/21166" target="_blank" rel="noopener noreferrer">PR #21166</a></td><td><a href="https://github.com/BerriAI/litellm/blob/main/litellm-proxy-extras/litellm_proxy_extras/migrations/20260213170952_access_group_change_to_model_name/migration.sql" target="_blank" rel="noopener noreferrer">Migration</a></td></tr><tr><td><code>LiteLLM_ManagedVectorStoreTable</code></td><td>New Table</td><td>Managed vector store tracking with model mappings</td><td>-</td><td><a href="https://github.com/BerriAI/litellm/blob/main/litellm-proxy-extras/litellm_proxy_extras/migrations/20260213105436_add_managed_vector_store_table/migration.sql" target="_blank" rel="noopener noreferrer">Migration</a></td></tr><tr><td><code>LiteLLM_TeamTable</code>, <code>LiteLLM_VerificationToken</code></td><td>New Column</td><td>Added <code>access_group_ids</code> text array</td><td><a href="https://github.com/BerriAI/litellm/pull/21022" target="_blank" rel="noopener noreferrer">PR #21022</a></td><td><a href="https://github.com/BerriAI/litellm/blob/main/litellm-proxy-extras/litellm_proxy_extras/migrations/20260212143306_add_access_group_table/migration.sql" target="_blank" rel="noopener noreferrer">Migration</a></td></tr><tr><td><code>LiteLLM_GuardrailsTable</code></td><td>New Column</td><td>Added <code>team_id</code> text column</td><td>-</td><td><a href="https://github.com/BerriAI/litellm/blob/main/litellm-proxy-extras/litellm_proxy_extras/migrations/20260214094754_schema_sync/migration.sql" target="_blank" rel="noopener noreferrer">Migration</a></td></tr></tbody></table>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates-14-updates">Documentation Updates (14 updates)<a href="https://docs.litellm.ai/release_notes/v1-81-12#documentation-updates-14-updates" class="hash-link" aria-label="Direct link to Documentation Updates (14 updates)" title="Direct link to Documentation Updates (14 updates)">​</a></h2>
<ul>
<li>LiteLLM Observatory section added to v1.81.9 release notes - <a href="https://github.com/BerriAI/litellm/pull/20675" target="_blank" rel="noopener noreferrer">PR #20675</a></li>
<li>Callback registration optimization added to release notes - <a href="https://github.com/BerriAI/litellm/pull/20681" target="_blank" rel="noopener noreferrer">PR #20681</a></li>
<li>Middleware performance blog post - <a href="https://github.com/BerriAI/litellm/pull/20677" target="_blank" rel="noopener noreferrer">PR #20677</a></li>
<li>UI Team Soft Budget documentation - <a href="https://github.com/BerriAI/litellm/pull/20669" target="_blank" rel="noopener noreferrer">PR #20669</a></li>
<li>UI Contributing and Troubleshooting guide - <a href="https://github.com/BerriAI/litellm/pull/20674" target="_blank" rel="noopener noreferrer">PR #20674</a></li>
<li>Reorganize Admin UI subsection - <a href="https://github.com/BerriAI/litellm/pull/20676" target="_blank" rel="noopener noreferrer">PR #20676</a></li>
<li>SDK proxy authentication (OAuth2/JWT auto-refresh) - <a href="https://github.com/BerriAI/litellm/pull/20680" target="_blank" rel="noopener noreferrer">PR #20680</a></li>
<li>Forward client headers to LLM API documentation fix - <a href="https://github.com/BerriAI/litellm/pull/20768" target="_blank" rel="noopener noreferrer">PR #20768</a></li>
<li>Add docs guide for using policies - <a href="https://github.com/BerriAI/litellm/pull/20914" target="_blank" rel="noopener noreferrer">PR #20914</a></li>
<li>Add native thinking param examples for Claude Opus 4.6 - <a href="https://github.com/BerriAI/litellm/pull/20799" target="_blank" rel="noopener noreferrer">PR #20799</a></li>
<li>Fix Claude Code MCP tutorial - <a href="https://github.com/BerriAI/litellm/pull/21145" target="_blank" rel="noopener noreferrer">PR #21145</a></li>
<li>Add API base URLs for Dashscope (International and China/Beijing) - <a href="https://github.com/BerriAI/litellm/pull/21083" target="_blank" rel="noopener noreferrer">PR #21083</a></li>
<li>Fix <code>DEFAULT_NUM_WORKERS_LITELLM_PROXY</code> default (1, not 4) - <a href="https://github.com/BerriAI/litellm/pull/21127" target="_blank" rel="noopener noreferrer">PR #21127</a></li>
<li>Correct ElevenLabs support status in README - <a href="https://github.com/BerriAI/litellm/pull/20643" target="_blank" rel="noopener noreferrer">PR #20643</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-81-12#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@iver56 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20643" target="_blank" rel="noopener noreferrer">PR #20643</a></li>
<li>@eliasaronson made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20666" target="_blank" rel="noopener noreferrer">PR #20666</a></li>
<li>@NirantK made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19656" target="_blank" rel="noopener noreferrer">PR #19656</a></li>
<li>@looksgood made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20919" target="_blank" rel="noopener noreferrer">PR #20919</a></li>
<li>@kelvin-tran made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20548" target="_blank" rel="noopener noreferrer">PR #20548</a></li>
<li>@bluet made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20873" target="_blank" rel="noopener noreferrer">PR #20873</a></li>
<li>@itayov made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20729" target="_blank" rel="noopener noreferrer">PR #20729</a></li>
<li>@CSteigstra made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20960" target="_blank" rel="noopener noreferrer">PR #20960</a></li>
<li>@rahulrd25 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20569" target="_blank" rel="noopener noreferrer">PR #20569</a></li>
<li>@muraliavarma made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20598" target="_blank" rel="noopener noreferrer">PR #20598</a></li>
<li>@joaokopernico made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21039" target="_blank" rel="noopener noreferrer">PR #21039</a></li>
<li>@datzscaler made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21077" target="_blank" rel="noopener noreferrer">PR #21077</a></li>
<li>@atapia27 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20922" target="_blank" rel="noopener noreferrer">PR #20922</a></li>
<li>@fpagny made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21121" target="_blank" rel="noopener noreferrer">PR #21121</a></li>
<li>@aidankovacic-8451 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/21119" target="_blank" rel="noopener noreferrer">PR #21119</a></li>
<li>@luisgallego-aily made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19935" target="_blank" rel="noopener noreferrer">PR #19935</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog">Full Changelog<a href="https://docs.litellm.ai/release_notes/v1-81-12#full-changelog" class="hash-link" aria-label="Direct link to Full Changelog" title="Direct link to Full Changelog">​</a></h2>
<p><a href="https://github.com/BerriAI/litellm/compare/v1.81.9.rc.1...v1.81.12.rc.1" target="_blank" rel="noopener noreferrer">v1.81.9.rc.1...v1.81.12.rc.1</a></p>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[v1.81.9 - Control which MCP Servers are exposed on the Internet]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-81-9</id>
        <link href="https://docs.litellm.ai/release_notes/v1-81-9"/>
        <updated>2026-02-07T00:00:00.000Z</updated>
        <summary type="html"><![CDATA[For each stable release, we now maintain a dedicated branch with the format litellmstablereleasebranchxxxxx for the version.]]></summary>
        <content type="html"><![CDATA[<div class="theme-admonition theme-admonition-info admonition_xJq3 alert alert--info"><div class="admonitionHeading_Gvgb"><span class="admonitionIcon_Rf37"><svg viewBox="0 0 14 16"><path fill-rule="evenodd" d="M7 2.3c3.14 0 5.7 2.56 5.7 5.7s-2.56 5.7-5.7 5.7A5.71 5.71 0 0 1 1.3 8c0-3.14 2.56-5.7 5.7-5.7zM7 1C3.14 1 0 4.14 0 8s3.14 7 7 7 7-3.14 7-7-3.14-7-7-7zm1 3H6v5h2V4zm0 6H6v2h2v-2z"></path></svg></span>Stable Release Branch</div><div class="admonitionContent_BuS1"><p>For each stable release, we now maintain a dedicated branch with the format <code>litellm_stable_release_branch_x_xx_xx</code> for the version.</p><p>This allows easier patching for day 0 model launches.</p><p><strong>Branch for v1.81.9:</strong> <a href="https://github.com/BerriAI/litellm/tree/litellm_stable_release_branch_1_81_9" target="_blank" rel="noopener noreferrer">litellm_stable_release_branch_1_81_9</a></p></div></div>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-81-9#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<!-- -->
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">ghcr.io/berriai/litellm:main-v1.81.9-stable</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.81.9</span><br></span></code></pre></div></div></div></div></div>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-81-9#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<ul>
<li><strong>Claude Opus 4.6</strong> - <a href="https://docs.litellm.ai/blog/claude_opus_4_6">Full support across Anthropic, AWS Bedrock, Azure AI, and Vertex AI with adaptive thinking and 1M context window</a></li>
<li><strong>A2A Agent Gateway</strong> - <a href="https://docs.litellm.ai/docs/a2a_invoking_agents">Call A2A (Agent-to-Agent) registered agents through the standard <code>/chat/completions</code> API</a></li>
<li><strong>Expose MCP servers on the public internet</strong> - <a href="https://docs.litellm.ai/docs/mcp_public_internet">Launch MCP servers with public/private visibility and IP-based access control for internet-facing deployments</a></li>
<li><strong>UI Team Soft Budget Alerts</strong> - <a href="https://docs.litellm.ai/docs/proxy/ui_team_soft_budget_alerts">Set soft budgets on teams and receive email alerts when spending crosses the threshold — without blocking requests</a></li>
<li><strong>Performance Optimizations</strong> - Multiple performance improvements including ~40% Prometheus CPU reduction, LRU caching, and optimized logging paths</li>
<li><strong>LiteLLM Observatory</strong> - <a href="https://docs.litellm.ai/blog/litellm-observatory">Automated 24-hour load tests</a></li>
<li><strong>30% Faster Request Processing for Callback-Heavy Deployments</strong> - [Performance improvement for callback heavy deployments]<a href="https://github.com/BerriAI/litellm/pull/20354" target="_blank" rel="noopener noreferrer">PR #20354</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="30-faster-request-processing-for-callback-heavy-deployments">30% Faster Request Processing for Callback-Heavy Deployments<a href="https://docs.litellm.ai/release_notes/v1-81-9#30-faster-request-processing-for-callback-heavy-deployments" class="hash-link" aria-label="Direct link to 30% Faster Request Processing for Callback-Heavy Deployments" title="Direct link to 30% Faster Request Processing for Callback-Heavy Deployments">​</a></h2>
<p>If you use logging callbacks like Langfuse, Datadog, or Prometheus, every request was paying an unnecessary cost: three loops that re-sorted your callbacks on every single request, even though the callback list hadn't changed. The more callbacks you had configured, the more time was wasted. We moved this work to happen once at startup instead of on every request. For deployments with the default callback set, this is a ~30% speedup in request setup. For deployments with many callbacks configured, the improvement is even larger.</p>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="litellm-observatory">LiteLLM Observatory<a href="https://docs.litellm.ai/release_notes/v1-81-9#litellm-observatory" class="hash-link" aria-label="Direct link to LiteLLM Observatory" title="Direct link to LiteLLM Observatory">​</a></h2>
<p>LiteLLM Observatory is a long-running release-validation system we built to catch regressions before they reach users. The system is built to be extensible—you can add new tests, configure models and failure thresholds, and queue runs against any deployment. Our goal is to achieve 100% coverage of LiteLLM functionality through these tests. We run 24-hour load tests against our production deployments before all releases, surfacing issues like resource lifecycle bugs, OOMs, and CPU regressions that only appear under sustained load.</p>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-servers-on-the-public-internet">MCP Servers on the Public Internet<a href="https://docs.litellm.ai/release_notes/v1-81-9#mcp-servers-on-the-public-internet" class="hash-link" aria-label="Direct link to MCP Servers on the Public Internet" title="Direct link to MCP Servers on the Public Internet">​</a></h2>
<p>This release makes it safe to expose MCP servers on the public internet by adding public/private visibility and IP-based access control. You can now run internet-facing MCP services while restricting access to trusted networks and keeping internal tools private.</p>
<p><a href="https://docs.litellm.ai/docs/mcp_public_internet">Get started</a></p>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAFCAYAAAB8ZH1oAAAACXBIWXMAACxLAAAsSwGlPZapAAAAe0lEQVR4nGXLywqDQBBE0fn/H4xCXEkCwczo+Ojpbh83mFWCBQeqFhXatqWqbjTNnbquObe7X4ScRx7PF12XyOOMFMN8w3z/sRHUVmI/k4aFWVaWsl2I7gT3lXdMxNQzThMigpmhql9nN1OCmRPTwJAnSlHO43HwlywHH6XNwQ6DbFoaAAAAAElFTkSuQmCC&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="334"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/mcp_internet.18c425e.640.png" srcset="/assets/ideal-img/mcp_internet.18c425e.640.png 640w,/assets/ideal-img/mcp_internet.93f669b.1920.png 1920w" width="640" height="334"></noscript></div>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="ui-team-soft-budget-alerts">UI Team Soft Budget Alerts<a href="https://docs.litellm.ai/release_notes/v1-81-9#ui-team-soft-budget-alerts" class="hash-link" aria-label="Direct link to UI Team Soft Budget Alerts" title="Direct link to UI Team Soft Budget Alerts">​</a></h2>
<p>Set a soft budget on any team to receive email alerts when spending crosses the threshold — without blocking any requests. Configure the threshold and alerting emails directly from the Admin UI, with no proxy restart needed.</p>
<p><a href="https://docs.litellm.ai/docs/proxy/ui_team_soft_budget_alerts">Get started</a></p>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAFCAYAAAB8ZH1oAAAACXBIWXMAACxLAAAsSwGlPZapAAAAi0lEQVR4nE2OuwrDMBRD/f+/FsjYDvXQEKgJFBqbOPa9fpwSD6UCLeJIyDyXBWstIQRUC61Vtk15OaXW+rOZ55lpmogxDhg6D5u43SP/Ms451nXlOA689yMUScQYSDmTpVALmFLKmM5XmNIArwvxPKmt0Vqnto4REVQVyZm3z7QORZXPviOqo9h75wumUcEi5RnllAAAAABJRU5ErkJggg==&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="334"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/ui_team_soft_budget_alerts.c3f20fa.640.png" srcset="/assets/ideal-img/ui_team_soft_budget_alerts.c3f20fa.640.png 640w,/assets/ideal-img/ui_team_soft_budget_alerts.89bac19.1920.png 1920w" width="640" height="334"></noscript></div>
<p>Let's dive in.</p>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-81-9#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support-13-new-models">New Model Support (13 new models)<a href="https://docs.litellm.ai/release_notes/v1-81-9#new-model-support-13-new-models" class="hash-link" aria-label="Direct link to New Model Support (13 new models)" title="Direct link to New Model Support (13 new models)">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th></tr></thead><tbody><tr><td>Anthropic</td><td><code>claude-opus-4-6</code></td><td>1M</td><td>$5.00</td><td>$25.00</td></tr><tr><td>AWS Bedrock</td><td><code>anthropic.claude-opus-4-6-v1</code></td><td>1M</td><td>$5.00</td><td>$25.00</td></tr><tr><td>Azure AI</td><td><code>azure_ai/claude-opus-4-6</code></td><td>200K</td><td>$5.00</td><td>$25.00</td></tr><tr><td>Vertex AI</td><td><code>vertex_ai/claude-opus-4-6</code></td><td>1M</td><td>$5.00</td><td>$25.00</td></tr><tr><td>Google Gemini</td><td><code>gemini/deep-research-pro-preview-12-2025</code></td><td>65K</td><td>$2.00</td><td>$12.00</td></tr><tr><td>Vertex AI</td><td><code>vertex_ai/deep-research-pro-preview-12-2025</code></td><td>65K</td><td>$2.00</td><td>$12.00</td></tr><tr><td>Moonshot</td><td><code>moonshot/kimi-k2.5</code></td><td>262K</td><td>$0.60</td><td>$3.00</td></tr><tr><td>OpenRouter</td><td><code>openrouter/qwen/qwen3-235b-a22b-2507</code></td><td>262K</td><td>$0.07</td><td>$0.10</td></tr><tr><td>OpenRouter</td><td><code>openrouter/qwen/qwen3-235b-a22b-thinking-2507</code></td><td>262K</td><td>$0.11</td><td>$0.60</td></tr><tr><td>Together AI</td><td><code>together_ai/zai-org/GLM-4.7</code></td><td>200K</td><td>$0.45</td><td>$2.00</td></tr><tr><td>Together AI</td><td><code>together_ai/moonshotai/Kimi-K2.5</code></td><td>256K</td><td>$0.50</td><td>$2.80</td></tr><tr><td>ElevenLabs</td><td><code>elevenlabs/eleven_v3</code></td><td>-</td><td>$0.18/1K chars</td><td>-</td></tr><tr><td>ElevenLabs</td><td><code>elevenlabs/eleven_multilingual_v2</code></td><td>-</td><td>$0.18/1K chars</td><td>-</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-81-9#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Full Claude Opus 4.6 support with adaptive thinking across all regions (us, eu, apac, au) - <a href="https://github.com/BerriAI/litellm/pull/20506" target="_blank" rel="noopener noreferrer">PR #20506</a>, <a href="https://github.com/BerriAI/litellm/pull/20508" target="_blank" rel="noopener noreferrer">PR #20508</a>, <a href="https://github.com/BerriAI/litellm/pull/20514" target="_blank" rel="noopener noreferrer">PR #20514</a>, <a href="https://github.com/BerriAI/litellm/pull/20551" target="_blank" rel="noopener noreferrer">PR #20551</a></li>
<li>Map reasoning content to anthropic thinking block (streaming + non-streaming) - <a href="https://github.com/BerriAI/litellm/pull/20254" target="_blank" rel="noopener noreferrer">PR #20254</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">AWS Bedrock</a></strong></p>
<ul>
<li>Add 1hr tiered caching costs for long-context models - <a href="https://github.com/BerriAI/litellm/pull/20214" target="_blank" rel="noopener noreferrer">PR #20214</a></li>
<li>Support TTL (1h) field in prompt caching for Bedrock Claude 4.5 models - <a href="https://github.com/BerriAI/litellm/pull/20338" target="_blank" rel="noopener noreferrer">PR #20338</a></li>
<li>Add Nova Sonic speech-to-speech model support - <a href="https://github.com/BerriAI/litellm/pull/20244" target="_blank" rel="noopener noreferrer">PR #20244</a></li>
<li>Fix empty assistant message for Converse API - <a href="https://github.com/BerriAI/litellm/pull/20390" target="_blank" rel="noopener noreferrer">PR #20390</a></li>
<li>Fix content blocked handling - <a href="https://github.com/BerriAI/litellm/pull/20606" target="_blank" rel="noopener noreferrer">PR #20606</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Google Gemini / Vertex AI</a></strong></p>
<ul>
<li>Add Gemini Deep Research model support - <a href="https://github.com/BerriAI/litellm/pull/20406" target="_blank" rel="noopener noreferrer">PR #20406</a></li>
<li>Fix Vertex AI Gemini streaming content_filter handling - <a href="https://github.com/BerriAI/litellm/pull/20105" target="_blank" rel="noopener noreferrer">PR #20105</a></li>
<li>Allow using OpenAI-style tools for <code>web_search</code> with Vertex AI/Gemini models - <a href="https://github.com/BerriAI/litellm/pull/20280" target="_blank" rel="noopener noreferrer">PR #20280</a></li>
<li>Fix <code>supports_native_streaming</code> for Gemini and Vertex AI models - <a href="https://github.com/BerriAI/litellm/pull/20408" target="_blank" rel="noopener noreferrer">PR #20408</a></li>
<li>Add mapping for responses tools in file IDs - <a href="https://github.com/BerriAI/litellm/pull/20402" target="_blank" rel="noopener noreferrer">PR #20402</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/cohere">Cohere</a></strong></p>
<ul>
<li>Support <code>dimensions</code> param for Cohere embed v4 - <a href="https://github.com/BerriAI/litellm/pull/20235" target="_blank" rel="noopener noreferrer">PR #20235</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/cerebras">Cerebras</a></strong></p>
<ul>
<li>Add reasoning param support for GPT OSS Cerebras - <a href="https://github.com/BerriAI/litellm/pull/20258" target="_blank" rel="noopener noreferrer">PR #20258</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/moonshot">Moonshot</a></strong></p>
<ul>
<li>Add Kimi K2.5 model entries - <a href="https://github.com/BerriAI/litellm/pull/20273" target="_blank" rel="noopener noreferrer">PR #20273</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openrouter">OpenRouter</a></strong></p>
<ul>
<li>Add Qwen3-235B models - <a href="https://github.com/BerriAI/litellm/pull/20455" target="_blank" rel="noopener noreferrer">PR #20455</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/togetherai">Together AI</a></strong></p>
<ul>
<li>Add GLM-4.7 and Kimi-K2.5 models - <a href="https://github.com/BerriAI/litellm/pull/20319" target="_blank" rel="noopener noreferrer">PR #20319</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/elevenlabs">ElevenLabs</a></strong></p>
<ul>
<li>Add <code>eleven_v3</code> and <code>eleven_multilingual_v2</code> TTS models - <a href="https://github.com/BerriAI/litellm/pull/20522" target="_blank" rel="noopener noreferrer">PR #20522</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vercel_ai_gateway">Vercel AI Gateway</a></strong></p>
<ul>
<li>Add missing capability flags to models - <a href="https://github.com/BerriAI/litellm/pull/20276" target="_blank" rel="noopener noreferrer">PR #20276</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/github_copilot">GitHub Copilot</a></strong></p>
<ul>
<li>Fix system prompts being dropped and auto-add required Copilot headers - <a href="https://github.com/BerriAI/litellm/pull/20113" target="_blank" rel="noopener noreferrer">PR #20113</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gigachat">GigaChat</a></strong></p>
<ul>
<li>Fix incorrect merging of consecutive user messages for GigaChat provider - <a href="https://github.com/BerriAI/litellm/pull/20341" target="_blank" rel="noopener noreferrer">PR #20341</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/xai_realtime">xAI</a></strong></p>
<ul>
<li>Add xAI <code>/realtime</code> API support - works with LiveKit SDK - <a href="https://github.com/BerriAI/litellm/pull/20381" target="_blank" rel="noopener noreferrer">PR #20381</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI</a></strong></p>
<ul>
<li>Add <code>gpt-5-search-api</code> model and docs clarifications - <a href="https://github.com/BerriAI/litellm/pull/20512" target="_blank" rel="noopener noreferrer">PR #20512</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-81-9#bug-fixes" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Fix extra inputs not permitted error for <code>provider_specific_fields</code> - <a href="https://github.com/BerriAI/litellm/pull/20334" target="_blank" rel="noopener noreferrer">PR #20334</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">AWS Bedrock</a></strong></p>
<ul>
<li>Fix: Managed Batches inconsistent state management for list and cancel batches - <a href="https://github.com/BerriAI/litellm/pull/20331" target="_blank" rel="noopener noreferrer">PR #20331</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI Embeddings</a></strong></p>
<ul>
<li>Fix <code>open_ai_embedding_models</code> to have <code>custom_llm_provider</code> None - <a href="https://github.com/BerriAI/litellm/pull/20253" target="_blank" rel="noopener noreferrer">PR #20253</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-81-9#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-81-9#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Messages API</a></strong></p>
<ul>
<li>Filter unsupported Claude Code beta headers for non-Anthropic providers - <a href="https://github.com/BerriAI/litellm/pull/20578" target="_blank" rel="noopener noreferrer">PR #20578</a></li>
<li>Fix inconsistent response format in <code>anthropic.messages.acreate()</code> when using non-Anthropic providers - <a href="https://github.com/BerriAI/litellm/pull/20442" target="_blank" rel="noopener noreferrer">PR #20442</a></li>
<li>Fix 404 on <code>/api/event_logging/batch</code> endpoint that caused Claude Code "route not found" errors - <a href="https://github.com/BerriAI/litellm/pull/20504" target="_blank" rel="noopener noreferrer">PR #20504</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/a2a">A2A Agent Gateway</a></strong></p>
<ul>
<li>Allow calling A2A agents through LiteLLM <code>/chat/completions</code> API - <a href="https://github.com/BerriAI/litellm/pull/20358" target="_blank" rel="noopener noreferrer">PR #20358</a></li>
<li>Use A2A registered agents with <code>/chat/completions</code> - <a href="https://github.com/BerriAI/litellm/pull/20362" target="_blank" rel="noopener noreferrer">PR #20362</a></li>
<li>Fix A2A agents deployed with localhost/internal URLs in their agent cards - <a href="https://github.com/BerriAI/litellm/pull/20604" target="_blank" rel="noopener noreferrer">PR #20604</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Files API</a></strong></p>
<ul>
<li>Add support for delete and GET via file_id for Gemini - <a href="https://github.com/BerriAI/litellm/pull/20329" target="_blank" rel="noopener noreferrer">PR #20329</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Add User-Agent customization support - <a href="https://github.com/BerriAI/litellm/pull/19881" target="_blank" rel="noopener noreferrer">PR #19881</a></li>
<li>Fix search tools not found when using per-request routers - <a href="https://github.com/BerriAI/litellm/pull/19818" target="_blank" rel="noopener noreferrer">PR #19818</a></li>
<li>Forward extra headers in chat - <a href="https://github.com/BerriAI/litellm/pull/20386" target="_blank" rel="noopener noreferrer">PR #20386</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-81-9#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-81-9#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong>SSO Configuration</strong></p>
<ul>
<li>SSO Config Team Mappings - <a href="https://github.com/BerriAI/litellm/pull/20111" target="_blank" rel="noopener noreferrer">PR #20111</a></li>
<li>UI - SSO: Add Team Mappings - <a href="https://github.com/BerriAI/litellm/pull/20299" target="_blank" rel="noopener noreferrer">PR #20299</a></li>
<li>Extract user roles from JWT access token for Keycloak compatibility - <a href="https://github.com/BerriAI/litellm/pull/20591" target="_blank" rel="noopener noreferrer">PR #20591</a></li>
</ul>
</li>
<li>
<p><strong>Auth / SDK</strong></p>
<ul>
<li>Add <code>proxy_auth</code> for auto OAuth2/JWT token management in SDK - <a href="https://github.com/BerriAI/litellm/pull/20238" target="_blank" rel="noopener noreferrer">PR #20238</a></li>
</ul>
</li>
<li>
<p><strong>Virtual Keys</strong></p>
<ul>
<li>Key <code>reset_spend</code> endpoint - <a href="https://github.com/BerriAI/litellm/pull/20305" target="_blank" rel="noopener noreferrer">PR #20305</a></li>
<li>UI - Keys: Allowed Routes to Key Info and Edit Pages - <a href="https://github.com/BerriAI/litellm/pull/20369" target="_blank" rel="noopener noreferrer">PR #20369</a></li>
<li>Add Key info endpoint object permission data - <a href="https://github.com/BerriAI/litellm/pull/20407" target="_blank" rel="noopener noreferrer">PR #20407</a></li>
<li>Keys and Teams Router Setting + Allow Override of Router Settings - <a href="https://github.com/BerriAI/litellm/pull/20205" target="_blank" rel="noopener noreferrer">PR #20205</a></li>
</ul>
</li>
<li>
<p><strong>Teams &amp; Budgets</strong></p>
<ul>
<li>Add <code>soft_budget</code> to Team Table + Create/Update Endpoints - <a href="https://github.com/BerriAI/litellm/pull/20530" target="_blank" rel="noopener noreferrer">PR #20530</a></li>
<li>Team Soft Budget Email Alerts - <a href="https://github.com/BerriAI/litellm/pull/20553" target="_blank" rel="noopener noreferrer">PR #20553</a></li>
<li>UI - Team Settings: Soft Budget + Alerting Emails - <a href="https://github.com/BerriAI/litellm/pull/20634" target="_blank" rel="noopener noreferrer">PR #20634</a></li>
<li>UI - User Budget Page: Unlimited Budget Checkbox - <a href="https://github.com/BerriAI/litellm/pull/20380" target="_blank" rel="noopener noreferrer">PR #20380</a></li>
<li><code>/user/update</code> allow for <code>max_budget</code> resets - <a href="https://github.com/BerriAI/litellm/pull/20375" target="_blank" rel="noopener noreferrer">PR #20375</a></li>
</ul>
</li>
<li>
<p><strong>UI Improvements</strong></p>
<ul>
<li>Default Team Settings: Migrate to use Reusable Model Select - <a href="https://github.com/BerriAI/litellm/pull/20310" target="_blank" rel="noopener noreferrer">PR #20310</a></li>
<li>Navbar: Option to Hide Community Engagement Buttons - <a href="https://github.com/BerriAI/litellm/pull/20308" target="_blank" rel="noopener noreferrer">PR #20308</a></li>
<li>Show team alias on Models health page - <a href="https://github.com/BerriAI/litellm/pull/20359" target="_blank" rel="noopener noreferrer">PR #20359</a></li>
<li>Admin Settings: Add option for Authentication for public AI Hub - <a href="https://github.com/BerriAI/litellm/pull/20444" target="_blank" rel="noopener noreferrer">PR #20444</a></li>
<li>Adjust daily spend date filtering for user timezone - <a href="https://github.com/BerriAI/litellm/pull/20472" target="_blank" rel="noopener noreferrer">PR #20472</a></li>
</ul>
</li>
<li>
<p><strong>SCIM</strong></p>
<ul>
<li>Add base <code>/scim/v2</code> endpoint for SCIM resource discovery - <a href="https://github.com/BerriAI/litellm/pull/20301" target="_blank" rel="noopener noreferrer">PR #20301</a></li>
</ul>
</li>
<li>
<p><strong>Proxy CLI</strong></p>
<ul>
<li>CLI arguments for RDS IAM auth - <a href="https://github.com/BerriAI/litellm/pull/20437" target="_blank" rel="noopener noreferrer">PR #20437</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs">Bugs<a href="https://docs.litellm.ai/release_notes/v1-81-9#bugs" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li>Fix: Remove unnecessary key blocking on UI login that prevented access - <a href="https://github.com/BerriAI/litellm/pull/20210" target="_blank" rel="noopener noreferrer">PR #20210</a></li>
<li>UI - Team Settings: Disable Global Guardrail Persistence - <a href="https://github.com/BerriAI/litellm/pull/20307" target="_blank" rel="noopener noreferrer">PR #20307</a></li>
<li>UI - Model Info Page: Fix Input and Output Labels - <a href="https://github.com/BerriAI/litellm/pull/20462" target="_blank" rel="noopener noreferrer">PR #20462</a></li>
<li>UI - Model Page: Column Resizing on Smaller Screens - <a href="https://github.com/BerriAI/litellm/pull/20599" target="_blank" rel="noopener noreferrer">PR #20599</a></li>
<li>Fix <code>/key/list</code> <code>user_id</code> Empty String Edge Case - <a href="https://github.com/BerriAI/litellm/pull/20623" target="_blank" rel="noopener noreferrer">PR #20623</a></li>
<li>Add array type checks for model, agent, and MCP hub data to prevent UI crashes - <a href="https://github.com/BerriAI/litellm/pull/20469" target="_blank" rel="noopener noreferrer">PR #20469</a></li>
<li>Fix unique constraint on daily tables + logging when updates fail - <a href="https://github.com/BerriAI/litellm/pull/20394" target="_blank" rel="noopener noreferrer">PR #20394</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="logging--guardrail--prompt-management-integrations">Logging / Guardrail / Prompt Management Integrations<a href="https://docs.litellm.ai/release_notes/v1-81-9#logging--guardrail--prompt-management-integrations" class="hash-link" aria-label="Direct link to Logging / Guardrail / Prompt Management Integrations" title="Direct link to Logging / Guardrail / Prompt Management Integrations">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes-3-fixes">Bug Fixes (3 fixes)<a href="https://docs.litellm.ai/release_notes/v1-81-9#bug-fixes-3-fixes" class="hash-link" aria-label="Direct link to Bug Fixes (3 fixes)" title="Direct link to Bug Fixes (3 fixes)">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#langfuse">Langfuse</a></strong></p>
<ul>
<li>Fix Langfuse OTEL trace export failing when spans contain null attributes - <a href="https://github.com/BerriAI/litellm/pull/20382" target="_blank" rel="noopener noreferrer">PR #20382</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#prometheus">Prometheus</a></strong></p>
<ul>
<li>Fix incorrect failure metrics labels causing miscounted error rates - <a href="https://github.com/BerriAI/litellm/pull/20152" target="_blank" rel="noopener noreferrer">PR #20152</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/alerting">Slack Alerts</a></strong></p>
<ul>
<li>Fix Slack alert delivery failing for certain budget threshold configurations - <a href="https://github.com/BerriAI/litellm/pull/20257" target="_blank" rel="noopener noreferrer">PR #20257</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails-7-updates">Guardrails (7 updates)<a href="https://docs.litellm.ai/release_notes/v1-81-9#guardrails-7-updates" class="hash-link" aria-label="Direct link to Guardrails (7 updates)" title="Direct link to Guardrails (7 updates)">​</a></h4>
<ul>
<li>
<p><strong>Custom Code Guardrails</strong></p>
<ul>
<li>Add HTTP support to custom code guardrails + Unified guardrails for MCP + Agent guardrail support - <a href="https://github.com/BerriAI/litellm/pull/20619" target="_blank" rel="noopener noreferrer">PR #20619</a></li>
<li>Custom Code Guardrails UI Playground - <a href="https://github.com/BerriAI/litellm/pull/20377" target="_blank" rel="noopener noreferrer">PR #20377</a></li>
</ul>
</li>
<li>
<p><strong>Team Bring-Your-Own Guardrails</strong></p>
<ul>
<li>Implement team-based isolation guardrails management - <a href="https://github.com/BerriAI/litellm/pull/20318" target="_blank" rel="noopener noreferrer">PR #20318</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/apply_guardrail">OpenAI Moderations</a></strong></p>
<ul>
<li>Ensure OpenAI Moderations Guard works with OpenAI Embeddings - <a href="https://github.com/BerriAI/litellm/pull/20523" target="_blank" rel="noopener noreferrer">PR #20523</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/apply_guardrail">GraySwan / Cygnal</a></strong></p>
<ul>
<li>Fix fail-open for GraySwan and pass metadata to Cygnal API endpoint - <a href="https://github.com/BerriAI/litellm/pull/19837" target="_blank" rel="noopener noreferrer">PR #19837</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Check for <code>model_response_choices</code> before guardrail input - <a href="https://github.com/BerriAI/litellm/pull/19784" target="_blank" rel="noopener noreferrer">PR #19784</a></li>
<li>Preserve streaming content on guardrail-sampled chunks - <a href="https://github.com/BerriAI/litellm/pull/20027" target="_blank" rel="noopener noreferrer">PR #20027</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-81-9#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li><strong>Support 0 cost models</strong> - Allow zero-cost model entries for internal/free-tier models - <a href="https://github.com/BerriAI/litellm/pull/20249" target="_blank" rel="noopener noreferrer">PR #20249</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-gateway-9-updates">MCP Gateway (9 updates)<a href="https://docs.litellm.ai/release_notes/v1-81-9#mcp-gateway-9-updates" class="hash-link" aria-label="Direct link to MCP Gateway (9 updates)" title="Direct link to MCP Gateway (9 updates)">​</a></h2>
<ul>
<li><strong>MCP Semantic Filtering</strong> - Filter MCP tools using semantic similarity to reduce tool sprawl for LLM calls - <a href="https://github.com/BerriAI/litellm/pull/20296" target="_blank" rel="noopener noreferrer">PR #20296</a>, <a href="https://github.com/BerriAI/litellm/pull/20316" target="_blank" rel="noopener noreferrer">PR #20316</a></li>
<li><strong>UI - MCP Semantic Filtering</strong> - Add support for MCP Semantic Filtering configuration on UI - <a href="https://github.com/BerriAI/litellm/pull/20454" target="_blank" rel="noopener noreferrer">PR #20454</a></li>
<li><strong>MCP IP-Based Access Control</strong> - Set MCP servers as private/public available on internet with IP-based restrictions - <a href="https://github.com/BerriAI/litellm/pull/20607" target="_blank" rel="noopener noreferrer">PR #20607</a>, <a href="https://github.com/BerriAI/litellm/pull/20620" target="_blank" rel="noopener noreferrer">PR #20620</a></li>
<li><strong>Fix MCP "Session not found" error</strong> on VSCode reconnect - <a href="https://github.com/BerriAI/litellm/pull/20298" target="_blank" rel="noopener noreferrer">PR #20298</a></li>
<li><strong>Fix OAuth2 'Capabilities: none' bug</strong> for upstream MCP servers - <a href="https://github.com/BerriAI/litellm/pull/20602" target="_blank" rel="noopener noreferrer">PR #20602</a></li>
<li><strong>Include Config Defined Search Tools</strong> in <code>/search_tools/list</code> - <a href="https://github.com/BerriAI/litellm/pull/20371" target="_blank" rel="noopener noreferrer">PR #20371</a></li>
<li><strong>UI - Search Tools</strong>: Show Config Defined Search Tools - <a href="https://github.com/BerriAI/litellm/pull/20436" target="_blank" rel="noopener noreferrer">PR #20436</a></li>
<li><strong>Ensure MCP permissions are enforced</strong> when using JWT Auth - <a href="https://github.com/BerriAI/litellm/pull/20383" target="_blank" rel="noopener noreferrer">PR #20383</a></li>
<li><strong>Fix <code>gcs_bucket_name</code> not being passed</strong> correctly for MCP server storage configuration - <a href="https://github.com/BerriAI/litellm/pull/20491" target="_blank" rel="noopener noreferrer">PR #20491</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements-14-improvements">Performance / Loadbalancing / Reliability improvements (14 improvements)<a href="https://docs.litellm.ai/release_notes/v1-81-9#performance--loadbalancing--reliability-improvements-14-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements (14 improvements)" title="Direct link to Performance / Loadbalancing / Reliability improvements (14 improvements)">​</a></h2>
<ul>
<li><strong>Prometheus ~40% CPU reduction</strong> - Parallelize budget metrics, fix caching bug, reduce CPU usage - <a href="https://github.com/BerriAI/litellm/pull/20544" target="_blank" rel="noopener noreferrer">PR #20544</a></li>
<li><strong>Prevent closed client errors</strong> by reverting httpx client caching - <a href="https://github.com/BerriAI/litellm/pull/20025" target="_blank" rel="noopener noreferrer">PR #20025</a></li>
<li><strong>Avoid unnecessary Router creation</strong> when no models or search tools are configured - <a href="https://github.com/BerriAI/litellm/pull/20661" target="_blank" rel="noopener noreferrer">PR #20661</a></li>
<li><strong>Optimize <code>wrapper_async</code></strong> with <code>CallTypes</code> caching and reduced lookups - <a href="https://github.com/BerriAI/litellm/pull/20204" target="_blank" rel="noopener noreferrer">PR #20204</a></li>
<li><strong>Cache <code>_get_relevant_args_to_use_for_logging()</code></strong> at module level - <a href="https://github.com/BerriAI/litellm/pull/20077" target="_blank" rel="noopener noreferrer">PR #20077</a></li>
<li><strong>LRU cache for <code>normalize_request_route</code></strong> - <a href="https://github.com/BerriAI/litellm/pull/19812" target="_blank" rel="noopener noreferrer">PR #19812</a></li>
<li><strong>Optimize <code>get_standard_logging_metadata</code></strong> with set intersection - <a href="https://github.com/BerriAI/litellm/pull/19685" target="_blank" rel="noopener noreferrer">PR #19685</a></li>
<li><strong>Early-exit guards in <code>completion_cost</code></strong> for unused features - <a href="https://github.com/BerriAI/litellm/pull/20020" target="_blank" rel="noopener noreferrer">PR #20020</a></li>
<li><strong>Optimize <code>get_litellm_params</code></strong> with sparse kwargs extraction - <a href="https://github.com/BerriAI/litellm/pull/19884" target="_blank" rel="noopener noreferrer">PR #19884</a></li>
<li><strong>Guard debug log f-strings</strong> and remove redundant dict copies - <a href="https://github.com/BerriAI/litellm/pull/19961" target="_blank" rel="noopener noreferrer">PR #19961</a></li>
<li><strong>Replace enum construction with frozenset lookup</strong> - <a href="https://github.com/BerriAI/litellm/pull/20302" target="_blank" rel="noopener noreferrer">PR #20302</a></li>
<li><strong>Guard debug f-string in <code>update_environment_variables</code></strong> - <a href="https://github.com/BerriAI/litellm/pull/20360" target="_blank" rel="noopener noreferrer">PR #20360</a></li>
<li><strong>Warn when budget lookup fails</strong> to surface silent caching misses - <a href="https://github.com/BerriAI/litellm/pull/20545" target="_blank" rel="noopener noreferrer">PR #20545</a></li>
<li><strong>Add INFO-level session reuse logging</strong> per request for better observability - <a href="https://github.com/BerriAI/litellm/pull/20597" target="_blank" rel="noopener noreferrer">PR #20597</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="database-changes">Database Changes<a href="https://docs.litellm.ai/release_notes/v1-81-9#database-changes" class="hash-link" aria-label="Direct link to Database Changes" title="Direct link to Database Changes">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="schema-updates">Schema Updates<a href="https://docs.litellm.ai/release_notes/v1-81-9#schema-updates" class="hash-link" aria-label="Direct link to Schema Updates" title="Direct link to Schema Updates">​</a></h3>
<table><thead><tr><th>Table</th><th>Change Type</th><th>Description</th><th>PR</th><th>Migration</th></tr></thead><tbody><tr><td><code>LiteLLM_TeamTable</code></td><td>New Column</td><td>Added <code>allow_team_guardrail_config</code> boolean field for team-based guardrail isolation</td><td><a href="https://github.com/BerriAI/litellm/pull/20318" target="_blank" rel="noopener noreferrer">PR #20318</a></td><td><a href="https://github.com/BerriAI/litellm/blob/main/litellm-proxy-extras/litellm_proxy_extras/migrations/20260205091235_allow_team_guardrail_config/migration.sql" target="_blank" rel="noopener noreferrer">Migration</a></td></tr><tr><td><code>LiteLLM_DeletedTeamTable</code></td><td>New Column</td><td>Added <code>allow_team_guardrail_config</code> boolean field</td><td><a href="https://github.com/BerriAI/litellm/pull/20318" target="_blank" rel="noopener noreferrer">PR #20318</a></td><td><a href="https://github.com/BerriAI/litellm/blob/main/litellm-proxy-extras/litellm_proxy_extras/migrations/20260205091235_allow_team_guardrail_config/migration.sql" target="_blank" rel="noopener noreferrer">Migration</a></td></tr><tr><td><code>LiteLLM_TeamTable</code></td><td>New Column</td><td>Added <code>soft_budget</code> (double precision) for soft budget alerting</td><td><a href="https://github.com/BerriAI/litellm/pull/20530" target="_blank" rel="noopener noreferrer">PR #20530</a></td><td><a href="https://github.com/BerriAI/litellm/blob/main/litellm-proxy-extras/litellm_proxy_extras/migrations/20260205144610_add_soft_budget_to_team_table/migration.sql" target="_blank" rel="noopener noreferrer">Migration</a></td></tr><tr><td><code>LiteLLM_DeletedTeamTable</code></td><td>New Column</td><td>Added <code>soft_budget</code> (double precision)</td><td><a href="https://github.com/BerriAI/litellm/pull/20653" target="_blank" rel="noopener noreferrer">PR #20653</a></td><td><a href="https://github.com/BerriAI/litellm/blob/main/litellm-proxy-extras/litellm_proxy_extras/migrations/20260207110613_add_soft_budget_to_deleted_teams_table/migration.sql" target="_blank" rel="noopener noreferrer">Migration</a></td></tr><tr><td><code>LiteLLM_MCPServerTable</code></td><td>New Column</td><td>Added <code>available_on_public_internet</code> boolean for MCP IP-based access control</td><td><a href="https://github.com/BerriAI/litellm/pull/20607" target="_blank" rel="noopener noreferrer">PR #20607</a></td><td><a href="https://github.com/BerriAI/litellm/blob/main/litellm-proxy-extras/litellm_proxy_extras/migrations/20260207093506_add_available_on_public_internet_to_mcp_servers/migration.sql" target="_blank" rel="noopener noreferrer">Migration</a></td></tr></tbody></table>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates-14-updates">Documentation Updates (14 updates)<a href="https://docs.litellm.ai/release_notes/v1-81-9#documentation-updates-14-updates" class="hash-link" aria-label="Direct link to Documentation Updates (14 updates)" title="Direct link to Documentation Updates (14 updates)">​</a></h2>
<ul>
<li>Add FAQ for setting up and verifying LITELLM_LICENSE - <a href="https://github.com/BerriAI/litellm/pull/20284" target="_blank" rel="noopener noreferrer">PR #20284</a></li>
<li>Model request tags documentation - <a href="https://github.com/BerriAI/litellm/pull/20290" target="_blank" rel="noopener noreferrer">PR #20290</a></li>
<li>Add Prisma migration troubleshooting guide - <a href="https://github.com/BerriAI/litellm/pull/20300" target="_blank" rel="noopener noreferrer">PR #20300</a></li>
<li>MCP Semantic Filtering documentation - <a href="https://github.com/BerriAI/litellm/pull/20316" target="_blank" rel="noopener noreferrer">PR #20316</a></li>
<li>Add CopilotKit SDK doc as supported agents SDK - <a href="https://github.com/BerriAI/litellm/pull/20396" target="_blank" rel="noopener noreferrer">PR #20396</a></li>
<li>Add documentation for Nova Sonic - <a href="https://github.com/BerriAI/litellm/pull/20320" target="_blank" rel="noopener noreferrer">PR #20320</a></li>
<li>Update Vertex AI Text to Speech doc to show use of audio - <a href="https://github.com/BerriAI/litellm/pull/20255" target="_blank" rel="noopener noreferrer">PR #20255</a></li>
<li>Improve Okta SSO setup guide with step-by-step instructions - <a href="https://github.com/BerriAI/litellm/pull/20353" target="_blank" rel="noopener noreferrer">PR #20353</a></li>
<li>Langfuse doc update - <a href="https://github.com/BerriAI/litellm/pull/20443" target="_blank" rel="noopener noreferrer">PR #20443</a></li>
<li>Expose MCPs on public internet documentation - <a href="https://github.com/BerriAI/litellm/pull/20626" target="_blank" rel="noopener noreferrer">PR #20626</a></li>
<li>Add blog post: Achieving Sub-Millisecond Proxy Overhead - <a href="https://github.com/BerriAI/litellm/pull/20309" target="_blank" rel="noopener noreferrer">PR #20309</a></li>
<li>Add blog post about litellm-observatory - <a href="https://github.com/BerriAI/litellm/pull/20622" target="_blank" rel="noopener noreferrer">PR #20622</a></li>
<li>Update Opus 4.6 blog with adaptive thinking - <a href="https://github.com/BerriAI/litellm/pull/20637" target="_blank" rel="noopener noreferrer">PR #20637</a></li>
<li><code>gpt-5-search-api</code> docs clarifications - <a href="https://github.com/BerriAI/litellm/pull/20512" target="_blank" rel="noopener noreferrer">PR #20512</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-81-9#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@Quentin-M made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19818" target="_blank" rel="noopener noreferrer">PR #19818</a></li>
<li>@amirzaushnizer made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20235" target="_blank" rel="noopener noreferrer">PR #20235</a></li>
<li>@cscguochang made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20214" target="_blank" rel="noopener noreferrer">PR #20214</a></li>
<li>@krauckbot made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20273" target="_blank" rel="noopener noreferrer">PR #20273</a></li>
<li>@agrattan0820 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19784" target="_blank" rel="noopener noreferrer">PR #19784</a></li>
<li>@nina-hu made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20472" target="_blank" rel="noopener noreferrer">PR #20472</a></li>
<li>@swayambhu94 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20469" target="_blank" rel="noopener noreferrer">PR #20469</a></li>
<li>@ssadedin made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20566" target="_blank" rel="noopener noreferrer">PR #20566</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog">Full Changelog<a href="https://docs.litellm.ai/release_notes/v1-81-9#full-changelog" class="hash-link" aria-label="Direct link to Full Changelog" title="Direct link to Full Changelog">​</a></h2>
<p><a href="https://github.com/BerriAI/litellm/compare/v1.81.6-nightly...v1.81.9" target="_blank" rel="noopener noreferrer">v1.81.6-nightly...v1.81.9</a></p>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[[Preview] v1.81.6 - Logs v2 with Tool Call Tracing]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-81-6</id>
        <link href="https://docs.litellm.ai/release_notes/v1-81-6"/>
        <updated>2026-01-31T00:00:00.000Z</updated>
        <summary type="html"><![CDATA[This release had known issues with CPU usage. This has been fixed in v1.81.9-stable.]]></summary>
        <content type="html"><![CDATA[<div class="theme-admonition theme-admonition-danger admonition_xJq3 alert alert--danger"><div class="admonitionHeading_Gvgb"><span class="admonitionIcon_Rf37"><svg viewBox="0 0 12 16"><path fill-rule="evenodd" d="M5.05.31c.81 2.17.41 3.38-.52 4.31C3.55 5.67 1.98 6.45.9 7.98c-1.45 2.05-1.7 6.53 3.53 7.7-2.2-1.16-2.67-4.52-.3-6.61-.61 2.03.53 3.33 1.94 2.86 1.39-.47 2.3.53 2.27 1.67-.02.78-.31 1.44-1.13 1.81 3.42-.59 4.78-3.42 4.78-5.56 0-2.84-2.53-3.22-1.25-5.61-1.52.13-2.03 1.13-1.89 2.75.09 1.08-1.02 1.8-1.86 1.33-.67-.41-.66-1.19-.06-1.78C8.18 5.31 8.68 2.45 5.05.32L5.03.3l.02.01z"></path></svg></span>Known Issue - CPU Usage</div><div class="admonitionContent_BuS1"><p>This release had known issues with CPU usage. This has been fixed in <a href="https://docs.litellm.ai/release_notes/v1-81-9">v1.81.9-stable</a>.</p><p><strong>We recommend using v1.81.9-stable instead.</strong></p></div></div>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-81-6#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<!-- -->
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-bash codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-bash codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">docker.litellm.ai/berriai/litellm:main-v1.81.6</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-bash codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-bash codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.81.6</span><br></span></code></pre></div></div></div></div></div>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-81-6#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<p>Logs View v2 with Tool Call Tracing - Redesigned logs interface with side panel, structured tool visualization, and error message search for faster debugging.</p>
<p>Let's dive in.</p>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="logs-view-v2-with-tool-call-tracing">Logs View v2 with Tool Call Tracing<a href="https://docs.litellm.ai/release_notes/v1-81-6#logs-view-v2-with-tool-call-tracing" class="hash-link" aria-label="Direct link to Logs View v2 with Tool Call Tracing" title="Direct link to Logs View v2 with Tool Call Tracing">​</a></h3>
<p>This release introduces comprehensive tool call tracing through LiteLLM's redesigned Logs View v2, enabling developers to debug and monitor AI agent workflows in production environments seamlessly.</p>
<p>This means you can now onboard use cases like tracing complex multi-step agent interactions, debugging tool execution failures, and monitoring MCP server calls while maintaining full visibility into request/response payloads with syntax highlighting.</p>
<p>Developers can access the new Logs View through LiteLLM's UI to inspect tool calls in structured format, search logs by error messages or request patterns, and correlate agent activities across sessions with collapsible side panel views.</p>
<!-- -->
<!-- -->
<p><a href="https://docs.litellm.ai/docs/proxy/ui_logs">Get Started</a></p>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-81-6#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support">New Model Support<a href="https://docs.litellm.ai/release_notes/v1-81-6#new-model-support" class="hash-link" aria-label="Direct link to New Model Support" title="Direct link to New Model Support">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th><th>Features</th></tr></thead><tbody><tr><td>AWS Bedrock</td><td><code>amazon.nova-2-pro-preview-20251202-v1:0</code></td><td>1M</td><td>$2.19</td><td>$17.50</td><td>Chat completions, vision, video, PDF, function calling, prompt caching, reasoning</td></tr><tr><td>Google Vertex AI</td><td><code>gemini-robotics-er-1.5-preview</code></td><td>1M</td><td>$0.30</td><td>$2.50</td><td>Chat completions, multimodal (text, image, video, audio), function calling, reasoning</td></tr><tr><td>OpenRouter</td><td><code>openrouter/xiaomi/mimo-v2-flash</code></td><td>262K</td><td>$0.09</td><td>$0.29</td><td>Chat completions, function calling, reasoning</td></tr><tr><td>OpenRouter</td><td><code>openrouter/moonshotai/kimi-k2.5</code></td><td>-</td><td>-</td><td>-</td><td>Chat completions</td></tr><tr><td>OpenRouter</td><td><code>openrouter/z-ai/glm-4.7</code></td><td>202K</td><td>$0.40</td><td>$1.50</td><td>Chat completions, vision, function calling, reasoning</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-81-6#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">AWS Bedrock</a></strong></p>
<ul>
<li>Messages API Bedrock Converse caching and PDF support - <a href="https://github.com/BerriAI/litellm/pull/19785" target="_blank" rel="noopener noreferrer">PR #19785</a></li>
<li>Translate advanced-tool-use to Bedrock-specific headers for Claude Opus 4.5 - <a href="https://github.com/BerriAI/litellm/pull/19841" target="_blank" rel="noopener noreferrer">PR #19841</a></li>
<li>Support tool search header translation for Sonnet 4.5 - <a href="https://github.com/BerriAI/litellm/pull/19871" target="_blank" rel="noopener noreferrer">PR #19871</a></li>
<li>Filter unsupported beta headers for AWS Bedrock Invoke API - <a href="https://github.com/BerriAI/litellm/pull/19877" target="_blank" rel="noopener noreferrer">PR #19877</a></li>
<li>Nova grounding improvements - <a href="https://github.com/BerriAI/litellm/pull/19598" target="_blank" rel="noopener noreferrer">PR #19598</a>, <a href="https://github.com/BerriAI/litellm/pull/20159" target="_blank" rel="noopener noreferrer">PR #20159</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Remove explicit cache_control null in tool_result content - <a href="https://github.com/BerriAI/litellm/pull/19919" target="_blank" rel="noopener noreferrer">PR #19919</a></li>
<li>Fix tool handling - <a href="https://github.com/BerriAI/litellm/pull/19805" target="_blank" rel="noopener noreferrer">PR #19805</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Google Gemini / Vertex AI</a></strong></p>
<ul>
<li>Add Gemini Robotics-ER 1.5 preview support - <a href="https://github.com/BerriAI/litellm/pull/19845" target="_blank" rel="noopener noreferrer">PR #19845</a></li>
<li>Support file retrieval in GoogleAIStudioFilesHandle - <a href="https://github.com/BerriAI/litellm/pull/20018" target="_blank" rel="noopener noreferrer">PR #20018</a></li>
<li>Add /delete endpoint support - <a href="https://github.com/BerriAI/litellm/pull/20055" target="_blank" rel="noopener noreferrer">PR #20055</a></li>
<li>Add custom_llm_provider as gemini translation - <a href="https://github.com/BerriAI/litellm/pull/19988" target="_blank" rel="noopener noreferrer">PR #19988</a></li>
<li>Subtract implicit cached tokens from text_tokens for correct cost calculation - <a href="https://github.com/BerriAI/litellm/pull/19775" target="_blank" rel="noopener noreferrer">PR #19775</a></li>
<li>Remove unsupported prompt-caching-scope-2026-01-05 header for vertex ai - <a href="https://github.com/BerriAI/litellm/pull/20058" target="_blank" rel="noopener noreferrer">PR #20058</a></li>
<li>Add disable flag for anthropic gemini cache translation - <a href="https://github.com/BerriAI/litellm/pull/20052" target="_blank" rel="noopener noreferrer">PR #20052</a></li>
<li>Convert image URLs to base64 in tool messages for Anthropic on Vertex AI - <a href="https://github.com/BerriAI/litellm/pull/19896" target="_blank" rel="noopener noreferrer">PR #19896</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/xai">xAI</a></strong></p>
<ul>
<li>Add grok reasoning content support - <a href="https://github.com/BerriAI/litellm/pull/19850" target="_blank" rel="noopener noreferrer">PR #19850</a></li>
<li>Add websearch params support for Responses API - <a href="https://github.com/BerriAI/litellm/pull/19915" target="_blank" rel="noopener noreferrer">PR #19915</a></li>
<li>Add routing of xai chat completions to responses when web search options is present - <a href="https://github.com/BerriAI/litellm/pull/20051" target="_blank" rel="noopener noreferrer">PR #20051</a></li>
<li>Correct cached token cost calculation - <a href="https://github.com/BerriAI/litellm/pull/19772" target="_blank" rel="noopener noreferrer">PR #19772</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/azure">Azure OpenAI</a></strong></p>
<ul>
<li>Use generic cost calculator for audio token pricing - <a href="https://github.com/BerriAI/litellm/pull/19771" target="_blank" rel="noopener noreferrer">PR #19771</a></li>
<li>Allow tool_choice for Azure GPT-5 chat models - <a href="https://github.com/BerriAI/litellm/pull/19813" target="_blank" rel="noopener noreferrer">PR #19813</a></li>
<li>Set gpt-5.2-codex mode to responses for Azure and OpenRouter - <a href="https://github.com/BerriAI/litellm/pull/19770" target="_blank" rel="noopener noreferrer">PR #19770</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI</a></strong></p>
<ul>
<li>Fix max_input_tokens for gpt-5.2-codex - <a href="https://github.com/BerriAI/litellm/pull/20009" target="_blank" rel="noopener noreferrer">PR #20009</a></li>
<li>Fix gpt-image-1.5 cost calculation not including output image tokens - <a href="https://github.com/BerriAI/litellm/pull/19515" target="_blank" rel="noopener noreferrer">PR #19515</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vllm">Hosted VLLM</a></strong></p>
<ul>
<li>Support thinking parameter in anthropic_messages() and .completion() - <a href="https://github.com/BerriAI/litellm/pull/19787" target="_blank" rel="noopener noreferrer">PR #19787</a></li>
<li>Route through base_llm_http_handler to support ssl_verify - <a href="https://github.com/BerriAI/litellm/pull/19893" target="_blank" rel="noopener noreferrer">PR #19893</a></li>
<li>Fix vllm embedding format - <a href="https://github.com/BerriAI/litellm/pull/20056" target="_blank" rel="noopener noreferrer">PR #20056</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/oci">OCI GenAI</a></strong></p>
<ul>
<li>Serialize imageUrl as object for OCI GenAI API - <a href="https://github.com/BerriAI/litellm/pull/19661" target="_blank" rel="noopener noreferrer">PR #19661</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/volcano">Volcengine</a></strong></p>
<ul>
<li>Add context for volcengine models (deepseek-v3-2, glm-4-7, kimi-k2-thinking) - <a href="https://github.com/BerriAI/litellm/pull/19335" target="_blank" rel="noopener noreferrer">PR #19335</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/">Chinese Providers</a></strong></p>
<ul>
<li>Add prompt caching and reasoning support for MiniMax, GLM, Xiaomi - <a href="https://github.com/BerriAI/litellm/pull/19924" target="_blank" rel="noopener noreferrer">PR #19924</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vercel_ai_gateway">Vercel AI Gateway</a></strong></p>
<ul>
<li>Add embeddings support - <a href="https://github.com/BerriAI/litellm/pull/19660" target="_blank" rel="noopener noreferrer">PR #19660</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-81-6#bug-fixes" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Google</a></strong></p>
<ul>
<li>Fix gemini-robotics-er-1.5-preview entry - <a href="https://github.com/BerriAI/litellm/pull/19974" target="_blank" rel="noopener noreferrer">PR #19974</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Fix output_tokens_details.reasoning_tokens None - <a href="https://github.com/BerriAI/litellm/pull/19914" target="_blank" rel="noopener noreferrer">PR #19914</a></li>
<li>Fix stream_chunk_builder to preserve images from streaming chunks - <a href="https://github.com/BerriAI/litellm/pull/19654" target="_blank" rel="noopener noreferrer">PR #19654</a></li>
<li>Fix aspectRatio mapping in image edit - <a href="https://github.com/BerriAI/litellm/pull/20053" target="_blank" rel="noopener noreferrer">PR #20053</a></li>
<li>Handle unknown models in Azure AI cost calculator - <a href="https://github.com/BerriAI/litellm/pull/20150" target="_blank" rel="noopener noreferrer">PR #20150</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gigachat">GigaChat</a></strong></p>
<ul>
<li>Ensure function content is valid JSON - <a href="https://github.com/BerriAI/litellm/pull/19232" target="_blank" rel="noopener noreferrer">PR #19232</a></li>
</ul>
</li>
</ul>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-81-6#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-81-6#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/mcp">Messages API (/messages)</a></strong></p>
<ul>
<li>Add LiteLLM x Claude Agent SDK Integration - <a href="https://github.com/BerriAI/litellm/pull/20035" target="_blank" rel="noopener noreferrer">PR #20035</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/mcp">A2A / MCP Gateway API (/a2a, /mcp)</a></strong></p>
<ul>
<li>Add A2A agent header-based context propagation support - <a href="https://github.com/BerriAI/litellm/pull/19504" target="_blank" rel="noopener noreferrer">PR #19504</a></li>
<li>Enable progress notifications for MCP tool calls - <a href="https://github.com/BerriAI/litellm/pull/19809" target="_blank" rel="noopener noreferrer">PR #19809</a></li>
<li>Fix support for non-standard MCP URL patterns - <a href="https://github.com/BerriAI/litellm/pull/19738" target="_blank" rel="noopener noreferrer">PR #19738</a></li>
<li>Add backward compatibility for legacy A2A card formats (/.well-known/agent.json) - <a href="https://github.com/BerriAI/litellm/pull/19949" target="_blank" rel="noopener noreferrer">PR #19949</a></li>
<li>Add support for agent parameter in /interactions endpoint - <a href="https://github.com/BerriAI/litellm/pull/19866" target="_blank" rel="noopener noreferrer">PR #19866</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/response_api">Responses API (/responses)</a></strong></p>
<ul>
<li>Fix custom_llm_provider for provider-specific params - <a href="https://github.com/BerriAI/litellm/pull/19798" target="_blank" rel="noopener noreferrer">PR #19798</a></li>
<li>Extract input tokens details as dict in ResponseAPILoggingUtils - <a href="https://github.com/BerriAI/litellm/pull/20046" target="_blank" rel="noopener noreferrer">PR #20046</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/batches">Batch API (/batches)</a></strong></p>
<ul>
<li>Fix /batches to return encoded ids (from managed objects table) - <a href="https://github.com/BerriAI/litellm/pull/19040" target="_blank" rel="noopener noreferrer">PR #19040</a></li>
<li>Fix Batch and File user level permissions - <a href="https://github.com/BerriAI/litellm/pull/19981" target="_blank" rel="noopener noreferrer">PR #19981</a></li>
<li>Add cost tracking and usage object in retrieve_batch call type - <a href="https://github.com/BerriAI/litellm/pull/19986" target="_blank" rel="noopener noreferrer">PR #19986</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/embedding/supported_embedding">Embeddings API (/embeddings)</a></strong></p>
<ul>
<li>Add supported input formats documentation - <a href="https://github.com/BerriAI/litellm/pull/20073" target="_blank" rel="noopener noreferrer">PR #20073</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/rag_ingest">RAG API (/rag/ingest, /vector_store)</a></strong></p>
<ul>
<li>Add UI for /rag/ingest API - Upload docs, pdfs etc to create vector stores - <a href="https://github.com/BerriAI/litellm/pull/19822" target="_blank" rel="noopener noreferrer">PR #19822</a></li>
<li>Add support for using S3 Vectors as Vector Store Provider - <a href="https://github.com/BerriAI/litellm/pull/19888" target="_blank" rel="noopener noreferrer">PR #19888</a></li>
<li>Add s3_vectors as provider on /vector_store/search API + UI for creating + PDF support - <a href="https://github.com/BerriAI/litellm/pull/19895" target="_blank" rel="noopener noreferrer">PR #19895</a></li>
<li>Add permission management for users and teams on Vector Stores - <a href="https://github.com/BerriAI/litellm/pull/19972" target="_blank" rel="noopener noreferrer">PR #19972</a></li>
<li>Enable router support for completions in RAG query pipeline - <a href="https://github.com/BerriAI/litellm/pull/19550" target="_blank" rel="noopener noreferrer">PR #19550</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/search">Search API (/search)</a></strong></p>
<ul>
<li>Add /list endpoint to list what search tools exist in router - <a href="https://github.com/BerriAI/litellm/pull/19969" target="_blank" rel="noopener noreferrer">PR #19969</a></li>
<li>Fix router search tools v2 integration - <a href="https://github.com/BerriAI/litellm/pull/19840" target="_blank" rel="noopener noreferrer">PR #19840</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/pass_through/intro">Passthrough Endpoints (/{provider}_passthrough)</a></strong></p>
<ul>
<li>Add /openai_passthrough route for OpenAI passthrough requests - <a href="https://github.com/BerriAI/litellm/pull/19989" target="_blank" rel="noopener noreferrer">PR #19989</a></li>
<li>Add support for configuring role_mappings via environment variables - <a href="https://github.com/BerriAI/litellm/pull/19498" target="_blank" rel="noopener noreferrer">PR #19498</a></li>
<li>Add Vertex AI LLM credentials sensitive keyword "vertex_credentials" for masking - <a href="https://github.com/BerriAI/litellm/pull/19551" target="_blank" rel="noopener noreferrer">PR #19551</a></li>
<li>Fix prevention of provider-prefixed model name leaks in responses - <a href="https://github.com/BerriAI/litellm/pull/19943" target="_blank" rel="noopener noreferrer">PR #19943</a></li>
<li>Fix proxy support for slashes in Google Vertex generateContent model names - <a href="https://github.com/BerriAI/litellm/pull/19737" target="_blank" rel="noopener noreferrer">PR #19737</a>, <a href="https://github.com/BerriAI/litellm/pull/19753" target="_blank" rel="noopener noreferrer">PR #19753</a></li>
<li>Support model names with slashes in Vertex AI passthrough URLs - <a href="https://github.com/BerriAI/litellm/pull/19944" target="_blank" rel="noopener noreferrer">PR #19944</a></li>
<li>Fix regression in Vertex AI passthroughs for router models - <a href="https://github.com/BerriAI/litellm/pull/19967" target="_blank" rel="noopener noreferrer">PR #19967</a></li>
<li>Add regression tests for Vertex AI passthrough model names - <a href="https://github.com/BerriAI/litellm/pull/19855" target="_blank" rel="noopener noreferrer">PR #19855</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs">Bugs<a href="https://docs.litellm.ai/release_notes/v1-81-6#bugs" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>General</strong>
<ul>
<li>Fix token calculations and refactor - <a href="https://github.com/BerriAI/litellm/pull/19696" target="_blank" rel="noopener noreferrer">PR #19696</a></li>
</ul>
</li>
</ul>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-81-6#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-81-6#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong>Proxy CLI Auth</strong></p>
<ul>
<li>Add configurable CLI JWT expiration via environment variable - <a href="https://github.com/BerriAI/litellm/pull/19780" target="_blank" rel="noopener noreferrer">PR #19780</a></li>
<li>Fix team cli auth flow - <a href="https://github.com/BerriAI/litellm/pull/19666" target="_blank" rel="noopener noreferrer">PR #19666</a></li>
</ul>
</li>
<li>
<p><strong>Virtual Keys</strong></p>
<ul>
<li>UI: Auto Truncation of Table Values - <a href="https://github.com/BerriAI/litellm/pull/19718" target="_blank" rel="noopener noreferrer">PR #19718</a></li>
<li>Fix Create Key: Expire Key Input Duration - <a href="https://github.com/BerriAI/litellm/pull/19807" target="_blank" rel="noopener noreferrer">PR #19807</a></li>
<li>Bulk Update Keys Endpoint - <a href="https://github.com/BerriAI/litellm/pull/19886" target="_blank" rel="noopener noreferrer">PR #19886</a></li>
</ul>
</li>
<li>
<p><strong>Logs View</strong></p>
<ul>
<li><strong>v2 Logs view with side panel and improved UX</strong> - <a href="https://github.com/BerriAI/litellm/pull/20091" target="_blank" rel="noopener noreferrer">PR #20091</a></li>
<li>New View to render "Tools" on Logs View - <a href="https://github.com/BerriAI/litellm/pull/20093" target="_blank" rel="noopener noreferrer">PR #20093</a></li>
<li>Add Pretty print view of request/response - <a href="https://github.com/BerriAI/litellm/pull/20096" target="_blank" rel="noopener noreferrer">PR #20096</a></li>
<li>Add error_message search in Spend Logs Endpoint - <a href="https://github.com/BerriAI/litellm/pull/19960" target="_blank" rel="noopener noreferrer">PR #19960</a></li>
<li>UI: Adding Error message search to ui spend logs - <a href="https://github.com/BerriAI/litellm/pull/19963" target="_blank" rel="noopener noreferrer">PR #19963</a></li>
<li>Spend Logs: Settings Modal - <a href="https://github.com/BerriAI/litellm/pull/19918" target="_blank" rel="noopener noreferrer">PR #19918</a></li>
<li>Fix error_code in Spend Logs metadata - <a href="https://github.com/BerriAI/litellm/pull/20015" target="_blank" rel="noopener noreferrer">PR #20015</a></li>
<li>Spend Logs: Show Current Store and Retention Status - <a href="https://github.com/BerriAI/litellm/pull/20017" target="_blank" rel="noopener noreferrer">PR #20017</a></li>
<li>Allow Dynamic Setting of store_prompts_in_spend_logs - <a href="https://github.com/BerriAI/litellm/pull/19913" target="_blank" rel="noopener noreferrer">PR #19913</a></li>
<li><a href="https://docs.litellm.ai/docs/proxy/ui_spend_log_settings">Docs: UI Spend Logs Settings</a> - <a href="https://github.com/BerriAI/litellm/pull/20197" target="_blank" rel="noopener noreferrer">PR #20197</a></li>
</ul>
</li>
<li>
<p><strong>Models + Endpoints</strong></p>
<ul>
<li>Add sortBy and sortOrder params for /v2/model/info - <a href="https://github.com/BerriAI/litellm/pull/19903" target="_blank" rel="noopener noreferrer">PR #19903</a></li>
<li>Fix Sorting for /v2/model/info - <a href="https://github.com/BerriAI/litellm/pull/19971" target="_blank" rel="noopener noreferrer">PR #19971</a></li>
<li>UI: Model Page Server Sort - <a href="https://github.com/BerriAI/litellm/pull/19908" target="_blank" rel="noopener noreferrer">PR #19908</a></li>
</ul>
</li>
<li>
<p><strong>Usage &amp; Analytics</strong></p>
<ul>
<li>UI: Usage Export: Breakdown by Teams and Keys - <a href="https://github.com/BerriAI/litellm/pull/19953" target="_blank" rel="noopener noreferrer">PR #19953</a></li>
<li>UI: Usage: Model Breakdown Per Key - <a href="https://github.com/BerriAI/litellm/pull/20039" target="_blank" rel="noopener noreferrer">PR #20039</a></li>
</ul>
</li>
<li>
<p><strong>UI Improvements</strong></p>
<ul>
<li>UI: Allow Admins to control what pages are visible on LeftNav - <a href="https://github.com/BerriAI/litellm/pull/19907" target="_blank" rel="noopener noreferrer">PR #19907</a></li>
<li>UI: Add Light/Dark Mode Switch for Development - <a href="https://github.com/BerriAI/litellm/pull/19804" target="_blank" rel="noopener noreferrer">PR #19804</a></li>
<li>UI: Dark Mode: Delete Resource Modal - <a href="https://github.com/BerriAI/litellm/pull/20098" target="_blank" rel="noopener noreferrer">PR #20098</a></li>
<li>UI: Tables: Reusable Table Sort Component - <a href="https://github.com/BerriAI/litellm/pull/19970" target="_blank" rel="noopener noreferrer">PR #19970</a></li>
<li>UI: New Badge Dot Render - <a href="https://github.com/BerriAI/litellm/pull/20024" target="_blank" rel="noopener noreferrer">PR #20024</a></li>
<li>UI: Feedback Prompts: Option To Hide Prompts - <a href="https://github.com/BerriAI/litellm/pull/19831" target="_blank" rel="noopener noreferrer">PR #19831</a></li>
<li>UI: Navbar: Fixed Default Logo + Bound Logo Box - <a href="https://github.com/BerriAI/litellm/pull/20092" target="_blank" rel="noopener noreferrer">PR #20092</a></li>
<li>UI: Navbar: User Dropdown - <a href="https://github.com/BerriAI/litellm/pull/20095" target="_blank" rel="noopener noreferrer">PR #20095</a></li>
<li>Change default key type from 'Default' to 'LLM API' - <a href="https://github.com/BerriAI/litellm/pull/19516" target="_blank" rel="noopener noreferrer">PR #19516</a></li>
</ul>
</li>
<li>
<p><strong>Team &amp; User Management</strong></p>
<ul>
<li>Fix /team/member_add User Email and ID Verifications - <a href="https://github.com/BerriAI/litellm/pull/19814" target="_blank" rel="noopener noreferrer">PR #19814</a></li>
<li>Fix SSO Email Case Sensitivity - <a href="https://github.com/BerriAI/litellm/pull/19799" target="_blank" rel="noopener noreferrer">PR #19799</a></li>
<li>UI: Internal User: Bulk Add - <a href="https://github.com/BerriAI/litellm/pull/19721" target="_blank" rel="noopener noreferrer">PR #19721</a></li>
</ul>
</li>
<li>
<p><strong>AI Gateway Features</strong></p>
<ul>
<li>Add support for making silent LLM calls without logging - <a href="https://github.com/BerriAI/litellm/pull/19544" target="_blank" rel="noopener noreferrer">PR #19544</a></li>
<li>UI: Fix MCP tools instructions to display comma-separated strings - <a href="https://github.com/BerriAI/litellm/pull/20101" target="_blank" rel="noopener noreferrer">PR #20101</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs-1">Bugs<a href="https://docs.litellm.ai/release_notes/v1-81-6#bugs-1" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li>Fix Model Name During Fallback - <a href="https://github.com/BerriAI/litellm/pull/20177" target="_blank" rel="noopener noreferrer">PR #20177</a></li>
<li>Fix Health Endpoints when Callback Objects Defined - <a href="https://github.com/BerriAI/litellm/pull/20182" target="_blank" rel="noopener noreferrer">PR #20182</a></li>
<li>Fix Unable to reset user max budget to unlimited - <a href="https://github.com/BerriAI/litellm/pull/19796" target="_blank" rel="noopener noreferrer">PR #19796</a></li>
<li>Fix Password comparison with non-ASCII characters - <a href="https://github.com/BerriAI/litellm/pull/19568" target="_blank" rel="noopener noreferrer">PR #19568</a></li>
<li>Correct error message for DISABLE_ADMIN_ENDPOINTS - <a href="https://github.com/BerriAI/litellm/pull/19861" target="_blank" rel="noopener noreferrer">PR #19861</a></li>
<li>Prevent clearing content filter patterns when editing guardrail - <a href="https://github.com/BerriAI/litellm/pull/19671" target="_blank" rel="noopener noreferrer">PR #19671</a></li>
<li>Fix Prompt Studio history to load tools and system messages - <a href="https://github.com/BerriAI/litellm/pull/19920" target="_blank" rel="noopener noreferrer">PR #19920</a></li>
<li>Add WATSONX_ZENAPIKEY to WatsonX credentials - <a href="https://github.com/BerriAI/litellm/pull/20086" target="_blank" rel="noopener noreferrer">PR #20086</a></li>
<li>UI: Vector Store: Allow Config Defined Models to Be Selected - <a href="https://github.com/BerriAI/litellm/pull/20031" target="_blank" rel="noopener noreferrer">PR #20031</a></li>
</ul>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="logging--guardrail--prompt-management-integrations">Logging / Guardrail / Prompt Management Integrations<a href="https://docs.litellm.ai/release_notes/v1-81-6#logging--guardrail--prompt-management-integrations" class="hash-link" aria-label="Direct link to Logging / Guardrail / Prompt Management Integrations" title="Direct link to Logging / Guardrail / Prompt Management Integrations">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-3">Features<a href="https://docs.litellm.ai/release_notes/v1-81-6#features-3" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#datadog">DataDog</a></strong></p>
<ul>
<li>Add agent support for LLM Observability - <a href="https://github.com/BerriAI/litellm/pull/19574" target="_blank" rel="noopener noreferrer">PR #19574</a></li>
<li>Add datadog cost management support and fix startup callback issue - <a href="https://github.com/BerriAI/litellm/pull/19584" target="_blank" rel="noopener noreferrer">PR #19584</a></li>
<li>Add datadog_llm_observability to /health/services allowed list - <a href="https://github.com/BerriAI/litellm/pull/19952" target="_blank" rel="noopener noreferrer">PR #19952</a></li>
<li>Check for agent mode before requiring DD_API_KEY/DD_SITE - <a href="https://github.com/BerriAI/litellm/pull/20156" target="_blank" rel="noopener noreferrer">PR #20156</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/observability/opentelemetry_integration">OpenTelemetry</a></strong></p>
<ul>
<li>Propagate JWT auth metadata to OTEL spans - <a href="https://github.com/BerriAI/litellm/pull/19627" target="_blank" rel="noopener noreferrer">PR #19627</a></li>
<li>Fix thread leak in dynamic header path - <a href="https://github.com/BerriAI/litellm/pull/19946" target="_blank" rel="noopener noreferrer">PR #19946</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#prometheus">Prometheus</a></strong></p>
<ul>
<li>Add callbacks and labels - <a href="https://github.com/BerriAI/litellm/pull/19708" target="_blank" rel="noopener noreferrer">PR #19708</a></li>
<li>Add clientip and user agent in metrics - <a href="https://github.com/BerriAI/litellm/pull/19717" target="_blank" rel="noopener noreferrer">PR #19717</a></li>
<li>Add tpm-rpm limit metrics - <a href="https://github.com/BerriAI/litellm/pull/19725" target="_blank" rel="noopener noreferrer">PR #19725</a></li>
<li>Add model_id label to metrics - <a href="https://github.com/BerriAI/litellm/pull/19678" target="_blank" rel="noopener noreferrer">PR #19678</a></li>
<li>Safely handle None metadata in logging - <a href="https://github.com/BerriAI/litellm/pull/19691" target="_blank" rel="noopener noreferrer">PR #19691</a></li>
<li>Resolve high CPU when router_settings in DB by avoiding REGISTRY.collect() - <a href="https://github.com/BerriAI/litellm/pull/20087" target="_blank" rel="noopener noreferrer">PR #20087</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#langfuse">Langfuse</a></strong></p>
<ul>
<li>Add litellm_callback_logging_failures_metric for Langfuse, Langfuse Otel and other Otel providers - <a href="https://github.com/BerriAI/litellm/pull/19636" target="_blank" rel="noopener noreferrer">PR #19636</a></li>
</ul>
</li>
<li>
<p><strong>General Logging</strong></p>
<ul>
<li>Use return value from CustomLogger.async_post_call_success_hook - <a href="https://github.com/BerriAI/litellm/pull/19670" target="_blank" rel="noopener noreferrer">PR #19670</a></li>
<li>Add async_post_call_response_headers_hook to CustomLogger - <a href="https://github.com/BerriAI/litellm/pull/20083" target="_blank" rel="noopener noreferrer">PR #20083</a></li>
<li>Add mock client factory pattern and mock support for PostHog, Helicone, and Braintrust integrations - <a href="https://github.com/BerriAI/litellm/pull/19707" target="_blank" rel="noopener noreferrer">PR #19707</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails">Guardrails<a href="https://docs.litellm.ai/release_notes/v1-81-6#guardrails" class="hash-link" aria-label="Direct link to Guardrails" title="Direct link to Guardrails">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/pii_masking_v2">Presidio</a></strong></p>
<ul>
<li>Reuse HTTP connections to prevent performance degradation - <a href="https://github.com/BerriAI/litellm/pull/19964" target="_blank" rel="noopener noreferrer">PR #19964</a></li>
</ul>
</li>
<li>
<p><strong>Onyx</strong></p>
<ul>
<li>Add timeout to onyx guardrail - <a href="https://github.com/BerriAI/litellm/pull/19731" target="_blank" rel="noopener noreferrer">PR #19731</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Add guardrail model argument feature - <a href="https://github.com/BerriAI/litellm/pull/19619" target="_blank" rel="noopener noreferrer">PR #19619</a></li>
<li>Fix guardrails issues with streaming-response regex - <a href="https://github.com/BerriAI/litellm/pull/19901" target="_blank" rel="noopener noreferrer">PR #19901</a></li>
<li>Remove enterprise requirement for guardrail monitoring (docs) - <a href="https://github.com/BerriAI/litellm/pull/19833" target="_blank" rel="noopener noreferrer">PR #19833</a></li>
</ul>
</li>
</ul>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-81-6#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li>Add event-driven coordination for global spend query to prevent cache stampede - <a href="https://github.com/BerriAI/litellm/pull/20030" target="_blank" rel="noopener noreferrer">PR #20030</a></li>
</ul>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements">Performance / Loadbalancing / Reliability improvements<a href="https://docs.litellm.ai/release_notes/v1-81-6#performance--loadbalancing--reliability-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements" title="Direct link to Performance / Loadbalancing / Reliability improvements">​</a></h2>
<ul>
<li><strong>Resolve high CPU when router_settings in DB</strong> - by avoiding REGISTRY.collect() in PrometheusServicesLogger - <a href="https://github.com/BerriAI/litellm/pull/20087" target="_blank" rel="noopener noreferrer">PR #20087</a></li>
<li><strong>Reuse HTTP connections in Presidio</strong> - to prevent performance degradation - <a href="https://github.com/BerriAI/litellm/pull/19964" target="_blank" rel="noopener noreferrer">PR #19964</a></li>
<li><strong>Event-driven coordination for global spend query</strong> - prevent cache stampede - <a href="https://github.com/BerriAI/litellm/pull/20030" target="_blank" rel="noopener noreferrer">PR #20030</a></li>
<li>Fix recursive Pydantic validation issue - <a href="https://github.com/BerriAI/litellm/pull/19531" target="_blank" rel="noopener noreferrer">PR #19531</a></li>
<li>Refactor argument handling into helper function to reduce code bloat - <a href="https://github.com/BerriAI/litellm/pull/19720" target="_blank" rel="noopener noreferrer">PR #19720</a></li>
<li>Optimize logo fetching and resolve MCP import blockers - <a href="https://github.com/BerriAI/litellm/pull/19719" target="_blank" rel="noopener noreferrer">PR #19719</a></li>
<li>Improve logo download performance using async HTTP client - <a href="https://github.com/BerriAI/litellm/pull/20155" target="_blank" rel="noopener noreferrer">PR #20155</a></li>
<li>Fix server root path configuration - <a href="https://github.com/BerriAI/litellm/pull/19790" target="_blank" rel="noopener noreferrer">PR #19790</a></li>
<li>Refactor: Extract transport context creation into separate method - <a href="https://github.com/BerriAI/litellm/pull/19794" target="_blank" rel="noopener noreferrer">PR #19794</a></li>
<li>Add native_background_mode configuration to override polling_via_cache for specific models - <a href="https://github.com/BerriAI/litellm/pull/19899" target="_blank" rel="noopener noreferrer">PR #19899</a></li>
<li>Initialize tiktoken environment at import time to enable offline usage - <a href="https://github.com/BerriAI/litellm/pull/19882" target="_blank" rel="noopener noreferrer">PR #19882</a></li>
<li>Improve tiktoken performance using local cache in lazy loading - <a href="https://github.com/BerriAI/litellm/pull/19774" target="_blank" rel="noopener noreferrer">PR #19774</a></li>
<li>Fix timeout errors in chat completion calls to be correctly reported in failure callbacks - <a href="https://github.com/BerriAI/litellm/pull/19842" target="_blank" rel="noopener noreferrer">PR #19842</a></li>
<li>Fix environment variable type handling for NUM_RETRIES - <a href="https://github.com/BerriAI/litellm/pull/19507" target="_blank" rel="noopener noreferrer">PR #19507</a></li>
<li>Use safe_deep_copy in silent experiment kwargs to prevent mutation - <a href="https://github.com/BerriAI/litellm/pull/20170" target="_blank" rel="noopener noreferrer">PR #20170</a></li>
<li>Improve error handling by inspecting BadRequestError after all other policy types - <a href="https://github.com/BerriAI/litellm/pull/19878" target="_blank" rel="noopener noreferrer">PR #19878</a></li>
</ul>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="database-changes">Database Changes<a href="https://docs.litellm.ai/release_notes/v1-81-6#database-changes" class="hash-link" aria-label="Direct link to Database Changes" title="Direct link to Database Changes">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="schema-updates">Schema Updates<a href="https://docs.litellm.ai/release_notes/v1-81-6#schema-updates" class="hash-link" aria-label="Direct link to Schema Updates" title="Direct link to Schema Updates">​</a></h3>
<table><thead><tr><th>Table</th><th>Change Type</th><th>Description</th><th>PR</th><th>Migration</th></tr></thead><tbody><tr><td><code>LiteLLM_ManagedVectorStoresTable</code></td><td>New Columns</td><td>Added <code>team_id</code> and <code>user_id</code> fields for permission management</td><td><a href="https://github.com/BerriAI/litellm/pull/19972" target="_blank" rel="noopener noreferrer">PR #19972</a></td><td><a href="https://github.com/BerriAI/litellm/blob/main/litellm-proxy-extras/litellm_proxy_extras/migrations/20260131150814_add_team_user_to_vector_stores/migration.sql" target="_blank" rel="noopener noreferrer">Migration</a></td></tr></tbody></table>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="migration-improvements">Migration Improvements<a href="https://docs.litellm.ai/release_notes/v1-81-6#migration-improvements" class="hash-link" aria-label="Direct link to Migration Improvements" title="Direct link to Migration Improvements">​</a></h3>
<ul>
<li>Fix Docker: Use correct schema path for Prisma generation - <a href="https://github.com/BerriAI/litellm/pull/19631" target="_blank" rel="noopener noreferrer">PR #19631</a></li>
<li>Resolve 'relation does not exist' migration errors in setup_database - <a href="https://github.com/BerriAI/litellm/pull/19281" target="_blank" rel="noopener noreferrer">PR #19281</a></li>
<li>Fix migration issue and improve Docker image stability - <a href="https://github.com/BerriAI/litellm/pull/19843" target="_blank" rel="noopener noreferrer">PR #19843</a></li>
<li>Run Prisma generate as nobody user in non-root Docker container for security - <a href="https://github.com/BerriAI/litellm/pull/20000" target="_blank" rel="noopener noreferrer">PR #20000</a></li>
<li>Bump litellm-proxy-extras version to 0.4.28 - <a href="https://github.com/BerriAI/litellm/pull/20166" target="_blank" rel="noopener noreferrer">PR #20166</a></li>
</ul>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates">Documentation Updates<a href="https://docs.litellm.ai/release_notes/v1-81-6#documentation-updates" class="hash-link" aria-label="Direct link to Documentation Updates" title="Direct link to Documentation Updates">​</a></h2>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/mcp">Add Claude Agents SDK x LiteLLM Guide</a></strong> - <a href="https://github.com/BerriAI/litellm/pull/20036" target="_blank" rel="noopener noreferrer">PR #20036</a></li>
<li><strong><a href="https://github.com/BerriAI/litellm/tree/main/cookbook" target="_blank" rel="noopener noreferrer">Add Cookbook: Using Claude Agent SDK + MCPs with LiteLLM</a></strong> - <a href="https://github.com/BerriAI/litellm/pull/20081" target="_blank" rel="noopener noreferrer">PR #20081</a></li>
<li>Fix A2A Python SDK URL in documentation - <a href="https://github.com/BerriAI/litellm/pull/19832" target="_blank" rel="noopener noreferrer">PR #19832</a></li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/sarvam">Add Sarvam usage documentation</a></strong> - <a href="https://github.com/BerriAI/litellm/pull/19844" target="_blank" rel="noopener noreferrer">PR #19844</a></li>
<li><strong><a href="https://docs.litellm.ai/docs/embedding/supported_embedding">Add supported input formats for embeddings</a></strong> - <a href="https://github.com/BerriAI/litellm/pull/20073" target="_blank" rel="noopener noreferrer">PR #20073</a></li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/ui_spend_log_settings">UI Spend Logs Settings Docs</a></strong> - <a href="https://github.com/BerriAI/litellm/pull/20197" target="_blank" rel="noopener noreferrer">PR #20197</a></li>
<li>Add OpenAI Agents SDK to OSS Adopters list in README - <a href="https://github.com/BerriAI/litellm/pull/19820" target="_blank" rel="noopener noreferrer">PR #19820</a></li>
<li>Update docs: Remove enterprise requirement for guardrail monitoring - <a href="https://github.com/BerriAI/litellm/pull/19833" target="_blank" rel="noopener noreferrer">PR #19833</a></li>
<li>Add missing environment variable documentation - <a href="https://github.com/BerriAI/litellm/pull/20138" target="_blank" rel="noopener noreferrer">PR #20138</a></li>
<li>Improve documentation blog index page - <a href="https://github.com/BerriAI/litellm/pull/20188" target="_blank" rel="noopener noreferrer">PR #20188</a></li>
</ul>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="infrastructure--testing-improvements">Infrastructure / Testing Improvements<a href="https://docs.litellm.ai/release_notes/v1-81-6#infrastructure--testing-improvements" class="hash-link" aria-label="Direct link to Infrastructure / Testing Improvements" title="Direct link to Infrastructure / Testing Improvements">​</a></h2>
<ul>
<li>Add test coverage for Router.get_valid_args and improve code coverage reporting - <a href="https://github.com/BerriAI/litellm/pull/19797" target="_blank" rel="noopener noreferrer">PR #19797</a></li>
<li>Add validation of model cost map as CI job - <a href="https://github.com/BerriAI/litellm/pull/19993" target="_blank" rel="noopener noreferrer">PR #19993</a></li>
<li>Add Realtime API benchmarks - <a href="https://github.com/BerriAI/litellm/pull/20074" target="_blank" rel="noopener noreferrer">PR #20074</a></li>
<li>Add Init Containers support in community helm chart - <a href="https://github.com/BerriAI/litellm/pull/19816" target="_blank" rel="noopener noreferrer">PR #19816</a></li>
<li>Add libsndfile to main Dockerfile for ARM64 audio processing support - <a href="https://github.com/BerriAI/litellm/pull/19776" target="_blank" rel="noopener noreferrer">PR #19776</a></li>
</ul>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-81-6#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@ruanjf made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19551" target="_blank" rel="noopener noreferrer">https://github.com/BerriAI/litellm/pull/19551</a></li>
<li>@moh-dev-stack made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19507" target="_blank" rel="noopener noreferrer">https://github.com/BerriAI/litellm/pull/19507</a></li>
<li>@formorter made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19498" target="_blank" rel="noopener noreferrer">https://github.com/BerriAI/litellm/pull/19498</a></li>
<li>@priyam-that made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19516" target="_blank" rel="noopener noreferrer">https://github.com/BerriAI/litellm/pull/19516</a></li>
<li>@marcosgriselli made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19550" target="_blank" rel="noopener noreferrer">https://github.com/BerriAI/litellm/pull/19550</a></li>
<li>@natimofeev made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19232" target="_blank" rel="noopener noreferrer">https://github.com/BerriAI/litellm/pull/19232</a></li>
<li>@zifeo made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19805" target="_blank" rel="noopener noreferrer">https://github.com/BerriAI/litellm/pull/19805</a></li>
<li>@pragyasardana made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19816" target="_blank" rel="noopener noreferrer">https://github.com/BerriAI/litellm/pull/19816</a></li>
<li>@ryewilson made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19833" target="_blank" rel="noopener noreferrer">https://github.com/BerriAI/litellm/pull/19833</a></li>
<li>@lizhen921 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19919" target="_blank" rel="noopener noreferrer">https://github.com/BerriAI/litellm/pull/19919</a></li>
<li>@boarder7395 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19666" target="_blank" rel="noopener noreferrer">https://github.com/BerriAI/litellm/pull/19666</a></li>
<li>@rushilchugh01 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19938" target="_blank" rel="noopener noreferrer">https://github.com/BerriAI/litellm/pull/19938</a></li>
<li>@cfchase made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19893" target="_blank" rel="noopener noreferrer">https://github.com/BerriAI/litellm/pull/19893</a></li>
<li>@ayim made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19872" target="_blank" rel="noopener noreferrer">https://github.com/BerriAI/litellm/pull/19872</a></li>
<li>@varunsripad123 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20018" target="_blank" rel="noopener noreferrer">https://github.com/BerriAI/litellm/pull/20018</a></li>
<li>@nht1206 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20046" target="_blank" rel="noopener noreferrer">https://github.com/BerriAI/litellm/pull/20046</a></li>
<li>@genga6 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/20009" target="_blank" rel="noopener noreferrer">https://github.com/BerriAI/litellm/pull/20009</a></li>
</ul>
<p><strong>Full Changelog</strong>: <a href="https://github.com/BerriAI/litellm/compare/v1.81.3.rc...v1.81.6" target="_blank" rel="noopener noreferrer">https://github.com/BerriAI/litellm/compare/v1.81.3.rc...v1.81.6</a></p>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[v1.81.3-stable - Performance - 25% CPU Usage Reduction]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-81-3</id>
        <link href="https://docs.litellm.ai/release_notes/v1-81-3"/>
        <updated>2026-01-26T10:00:00.000Z</updated>
        <summary type="html"><![CDATA[Deploy this version]]></summary>
        <content type="html"><![CDATA[<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-81-3#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">docker.litellm.ai/berriai/litellm:v1.81.3-stable</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.81.3.rc.2</span><br></span></code></pre></div></div></div></div></div>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-81-3#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support">New Model Support<a href="https://docs.litellm.ai/release_notes/v1-81-3#new-model-support" class="hash-link" aria-label="Direct link to New Model Support" title="Direct link to New Model Support">​</a></h3>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th><th>Deprecation Date</th></tr></thead><tbody><tr><td>OpenAI</td><td><code>gpt-audio</code>, <code>gpt-audio-2025-08-28</code></td><td>128K</td><td>$32/1M audio tokens, $2.5/1M text tokens</td><td>$64/1M audio tokens, $10/1M text tokens</td><td>-</td></tr><tr><td>OpenAI</td><td><code>gpt-audio-mini</code>, <code>gpt-audio-mini-2025-08-28</code></td><td>128K</td><td>$10/1M audio tokens, $0.6/1M text tokens</td><td>$20/1M audio tokens, $2.4/1M text tokens</td><td>-</td></tr><tr><td>Deepinfra, Vertex AI, Google AI Studio, OpenRouter, Vercel AI Gateway</td><td><code>gemini-2.0-flash-001</code>, <code>gemini-2.0-flash</code></td><td>-</td><td>-</td><td>-</td><td>2026-03-31</td></tr><tr><td>Groq</td><td><code>openai/gpt-oss-120b</code></td><td>131K</td><td>0.075/1M cache read</td><td>0.6/1M output tokens</td><td>-</td></tr><tr><td>Groq</td><td><code>groq/openai/gpt-oss-20b</code></td><td>131K</td><td>0.0375/1M cache read, $0.075/1M text tokens</td><td>0.3/1M output tokens</td><td>-</td></tr><tr><td>Vertex AI</td><td><code>gemini-2.5-computer-use-preview-10-2025</code></td><td>128K</td><td>$1.25</td><td>$10</td><td>-</td></tr><tr><td>Azure AI</td><td><code>claude-haiku-4-5</code></td><td>$1.25/1M cache read, $2/1M cache read above 1 hr, $0.1/1M text tokens</td><td>$5/1M output tokens</td><td>-</td><td></td></tr><tr><td>Azure AI</td><td><code>claude-sonnet-4-5</code></td><td>$3.75/1M cache read, $6/1M cache read above 1 hr, $3/1M text tokens</td><td>$15/1M output tokens</td><td>-</td><td></td></tr><tr><td>Azure AI</td><td><code>claude-opus-4-5</code></td><td>$6.25/1M cache read, $10/1M cache read above 1 hr, $0.5/1M text tokens</td><td>$25/1M output tokens</td><td>-</td><td></td></tr><tr><td>Azure AI</td><td><code>claude-opus-4-1</code></td><td>$18.75/1M cache read, $30/1M cache read above 1 hr, $1.5/1M text tokens</td><td>$75/1M output tokens</td><td>-</td><td></td></tr></tbody></table>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-81-3#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI</a></strong></p>
<ul>
<li>Add gpt-audio and gpt-audio-mini models to pricing - <a href="https://github.com/BerriAI/litellm/pull/19509" target="_blank" rel="noopener noreferrer">PR #19509</a></li>
<li>correct audio token costs for gpt-4o-audio-preview models - <a href="https://github.com/BerriAI/litellm/pull/19500" target="_blank" rel="noopener noreferrer">PR #19500</a></li>
<li>Limit stop sequence as per openai spec (ensures JetBrains IDE compatibility) - <a href="https://github.com/BerriAI/litellm/pull/19562" target="_blank" rel="noopener noreferrer">PR #19562</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vertex">VertexAI</a></strong></p>
<ul>
<li>Docs - Google Workload Identity Federation (WIF) support - <a href="https://github.com/BerriAI/litellm/pull/19320" target="_blank" rel="noopener noreferrer">PR #19320</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock_agentcore">Agentcore</a></strong></p>
<ul>
<li>Fixes streaming issues with AWS Bedrock AgentCore where responses would stop after the first chunk, particularly affecting OAuth-enabled agents - <a href="https://github.com/BerriAI/litellm/pull/17141" target="_blank" rel="noopener noreferrer">PR #17141</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/chatgpt">Chatgpt</a></strong></p>
<ul>
<li>Adds support for calling chatgpt subscription via LiteLLM - <a href="https://github.com/BerriAI/litellm/pull/19030" target="_blank" rel="noopener noreferrer">PR #19030</a></li>
<li>Adds responses API bridge support for chatgpt subscription provider - <a href="https://github.com/BerriAI/litellm/pull/19030" target="_blank" rel="noopener noreferrer">PR #19030</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong></p>
<ul>
<li>support for output format for bedrock invoke via v1/messages - <a href="https://github.com/BerriAI/litellm/pull/19560" target="_blank" rel="noopener noreferrer">PR #19560</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/azure/azure">Azure</a></strong></p>
<ul>
<li>Add support for Azure OpenAI v1 API - <a href="https://github.com/BerriAI/litellm/pull/19313" target="_blank" rel="noopener noreferrer">PR #19313</a></li>
<li>preserve content_policy_violation details for images (#19328) - <a href="https://github.com/BerriAI/litellm/pull/19372" target="_blank" rel="noopener noreferrer">PR #19372</a></li>
<li>Support OpenAI-format nested tool definitions for Responses API - <a href="https://github.com/BerriAI/litellm/pull/19526" target="_blank" rel="noopener noreferrer">PR #19526</a></li>
</ul>
</li>
<li>
<p><strong>Gemini(<a href="https://docs.litellm.ai/docs/providers/vertex">Vertex AI</a>, <a href="https://docs.litellm.ai/docs/providers/gemini">Google AI Studio</a>)</strong></p>
<ul>
<li>use responseJsonSchema for Gemini 2.0+ models - <a href="https://github.com/BerriAI/litellm/pull/19314" target="_blank" rel="noopener noreferrer">PR #19314</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/volcano">Volcengine</a></strong></p>
<ul>
<li>Support Volcengine responses api - <a href="https://github.com/BerriAI/litellm/pull/18508" target="_blank" rel="noopener noreferrer">PR #18508</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Add Support for calling Claude Code Max subscriptions via LiteLLM - <a href="https://github.com/BerriAI/litellm/pull/19453" target="_blank" rel="noopener noreferrer">PR #19453</a></li>
<li>Add Structured output for /v1/messages with Anthropic API, Azure Anthropic API, Bedrock Converse - <a href="https://github.com/BerriAI/litellm/pull/19545" target="_blank" rel="noopener noreferrer">PR #19545</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/search/brave">Brave Search</a></strong></p>
<ul>
<li>New Search provider - <a href="https://github.com/BerriAI/litellm/pull/19433" target="_blank" rel="noopener noreferrer">PR #19433</a></li>
</ul>
</li>
<li>
<p><strong>Sarvam ai</strong></p>
<ul>
<li>Add support for new sarvam models  - <a href="https://github.com/BerriAI/litellm/pull/19479" target="_blank" rel="noopener noreferrer">PR #19479</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gmi">GMI</a></strong></p>
<ul>
<li>add GMI Cloud provider support - <a href="https://github.com/BerriAI/litellm/pull/19376" target="_blank" rel="noopener noreferrer">PR #19376</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-81-3#bug-fixes" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Fix anthropic-beta sent client side being overridden instead of appended to - <a href="https://github.com/BerriAI/litellm/pull/19343" target="_blank" rel="noopener noreferrer">PR #19343</a></li>
<li>Filter out unsupported fields from JSON schema for Anthropic's output_format API - <a href="https://github.com/BerriAI/litellm/pull/19482" target="_blank" rel="noopener noreferrer">PR #19482</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong></p>
<ul>
<li>Expose stability models via /image_edits endpoint and ensure proper request transformation - <a href="https://github.com/BerriAI/litellm/pull/19323" target="_blank" rel="noopener noreferrer">PR #19323</a></li>
<li>Claude Code x Bedrock Invoke fails with advanced-tool-use-2025-11-20 - <a href="https://github.com/BerriAI/litellm/pull/19373" target="_blank" rel="noopener noreferrer">PR #19373</a></li>
<li>deduplicate tool calls in assistant history - <a href="https://github.com/BerriAI/litellm/pull/19324" target="_blank" rel="noopener noreferrer">PR #19324</a></li>
<li>fix: correct us.anthropic.claude-opus-4-5 In-region pricing - <a href="https://github.com/BerriAI/litellm/pull/19310" target="_blank" rel="noopener noreferrer">PR #19310</a></li>
<li>Fix request validation errors when using Claude 4 via bedrock invoke - <a href="https://github.com/BerriAI/litellm/pull/19381" target="_blank" rel="noopener noreferrer">PR #19381</a></li>
<li>Handle thinking with tool calls for Claude 4 models - <a href="https://github.com/BerriAI/litellm/pull/19506" target="_blank" rel="noopener noreferrer">PR #19506</a></li>
<li>correct streaming choice index for tool calls - <a href="https://github.com/BerriAI/litellm/pull/19506" target="_blank" rel="noopener noreferrer">PR #19506</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/ollama">Ollama</a></strong></p>
<ul>
<li>Fix tool call errors due with improved message extraction - <a href="https://github.com/BerriAI/litellm/pull/19369" target="_blank" rel="noopener noreferrer">PR #19369</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vertex">VertexAI</a></strong></p>
<ul>
<li>Removed optional&nbsp;vertex_count_tokens_location&nbsp;param before request is sent to vertex - <a href="https://github.com/BerriAI/litellm/pull/19359" target="_blank" rel="noopener noreferrer">PR #19359</a></li>
</ul>
</li>
<li>
<p><strong>Gemini(<a href="https://docs.litellm.ai/docs/providers/vertex">Vertex AI</a>, <a href="https://docs.litellm.ai/docs/providers/gemini">Google AI Studio</a>)</strong></p>
<ul>
<li>Supports setting media_resolution and fps parameters on each video file, when using Gemini video understanding - <a href="https://github.com/BerriAI/litellm/pull/19273" target="_blank" rel="noopener noreferrer">PR #19273</a></li>
<li>handle reasoning_effort as dict from OpenAI Agents SDK - <a href="https://github.com/BerriAI/litellm/pull/19419" target="_blank" rel="noopener noreferrer">PR #19419</a></li>
<li>add file content support in tool results - <a href="https://github.com/BerriAI/litellm/pull/19416" target="_blank" rel="noopener noreferrer">PR #19416</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/azure_ai">Azure</a></strong></p>
<ul>
<li>Fix Azure AI costs for Anthropic models - <a href="https://github.com/BerriAI/litellm/pull/19530" target="_blank" rel="noopener noreferrer">PR #19530</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gigachat">Giga Chat</a></strong></p>
<ul>
<li>Add tool choice mapping - <a href="https://github.com/BerriAI/litellm/pull/19645" target="_blank" rel="noopener noreferrer">PR #19645</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="ai-api-endpoints-llms-mcp-agents">AI API Endpoints (LLMs, MCP, Agents)<a href="https://docs.litellm.ai/release_notes/v1-81-3#ai-api-endpoints-llms-mcp-agents" class="hash-link" aria-label="Direct link to AI API Endpoints (LLMs, MCP, Agents)" title="Direct link to AI API Endpoints (LLMs, MCP, Agents)">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-81-3#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/files_endpoints">Files API</a></strong></p>
<ul>
<li>Add managed files support when load_balancing is True - <a href="https://github.com/BerriAI/litellm/pull/19338" target="_blank" rel="noopener noreferrer">PR #19338</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/tutorials/claude_code_plugin_marketplace">Claude Plugin Marketplace</a></strong></p>
<ul>
<li>Add self hosted Claude Code Plugin Marketplace - <a href="https://github.com/BerriAI/litellm/pull/19378" target="_blank" rel="noopener noreferrer">PR #19378</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/mcp">MCP</a></strong></p>
<ul>
<li>Add MCP Protocol version&nbsp;2025-11-25&nbsp;support - <a href="https://github.com/BerriAI/litellm/pull/19379" target="_blank" rel="noopener noreferrer">PR #19379</a></li>
<li>Log MCP tool calls and list tools in the LiteLLM Spend Logs table for easier debugging - <a href="https://github.com/BerriAI/litellm/pull/19469" target="_blank" rel="noopener noreferrer">PR #19469</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vertex">Vertex AI</a></strong></p>
<ul>
<li>Ensure only anthropic betas are forwarded down to LLM API (by default) - <a href="https://github.com/BerriAI/litellm/pull/19542" target="_blank" rel="noopener noreferrer">PR #19542</a></li>
<li>Allow overriding to support forwarding incoming headers are forwarded down to target - <a href="https://github.com/BerriAI/litellm/pull/19524" target="_blank" rel="noopener noreferrer">PR #19524</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/completion/input">Chat/Completions</a></strong></p>
<ul>
<li>Add MCP tools response to chat completions - <a href="https://github.com/BerriAI/litellm/pull/19552" target="_blank" rel="noopener noreferrer">PR #19552</a></li>
<li>Add custom vertex ai finish reasons to the output - <a href="https://github.com/BerriAI/litellm/pull/19558" target="_blank" rel="noopener noreferrer">PR #19558</a></li>
<li>Return MCP execution in /chat/completions before model output during streaming - <a href="https://github.com/BerriAI/litellm/pull/19623" target="_blank" rel="noopener noreferrer">PR #19623</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bugs">Bugs<a href="https://docs.litellm.ai/release_notes/v1-81-3#bugs" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/response_api">Responses API</a></strong></p>
<ul>
<li>Fix duplicate messages during MCP streaming tool execution - <a href="https://github.com/BerriAI/litellm/pull/19317" target="_blank" rel="noopener noreferrer">PR #19317</a></li>
<li>Fix pickle error when using OpenAI's&nbsp;Responses API&nbsp;with&nbsp;stream=True&nbsp;and&nbsp;tool_choice&nbsp;of type&nbsp;allowed_tools&nbsp;(an&nbsp;OpenAI-native parameter) - <a href="https://github.com/BerriAI/litellm/pull/17205" target="_blank" rel="noopener noreferrer">PR #17205</a></li>
<li>stream tool call events for non-openai models - <a href="https://github.com/BerriAI/litellm/pull/19368" target="_blank" rel="noopener noreferrer">PR #19368</a></li>
<li>preserve tool output ordering for gemini in responses bridge&nbsp;- <a href="https://github.com/BerriAI/litellm/pull/19360" target="_blank" rel="noopener noreferrer">PR #19360</a></li>
<li>Add ID caching to prevent ID mismatch text-start and text-delta - <a href="https://github.com/BerriAI/litellm/pull/19390" target="_blank" rel="noopener noreferrer">PR #19390</a></li>
<li>Include output_item, reasoning_summary_Text_done and reasoning_summary_part_done events for non-openai models - <a href="https://github.com/BerriAI/litellm/pull/19472" target="_blank" rel="noopener noreferrer">PR #19472</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/completion/input">Chat/Completions</a></strong></p>
<ul>
<li>fix: drop_params not dropping prompt_cache_key for non-OpenAI providers - <a href="https://github.com/BerriAI/litellm/pull/19346" target="_blank" rel="noopener noreferrer">PR #19346</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/realtime">Realtime API</a></strong></p>
<ul>
<li>disable SSL for ws:// WebSocket connections - <a href="https://github.com/BerriAI/litellm/pull/19345" target="_blank" rel="noopener noreferrer">PR #19345</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/generateContent">Generate Content</a></strong></p>
<ul>
<li>Log actual user input when google genai/vertex endpoints are called client-side - <a href="https://github.com/BerriAI/litellm/pull/19156" target="_blank" rel="noopener noreferrer">PR #19156</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/anthropic_count_tokens">/messages/count_tokens Anthropic Token Counting</a></strong></p>
<ul>
<li>ensure it works for Anthropic, Azure AI Anthropic on AI Gateway - <a href="https://github.com/BerriAI/litellm/pull/19432" target="_blank" rel="noopener noreferrer">PR #19432</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/mcp">MCP</a></strong></p>
<ul>
<li>forward static_headers to MCP servers - <a href="https://github.com/BerriAI/litellm/pull/19366" target="_blank" rel="noopener noreferrer">PR #19366</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/batches">Batch API</a></strong></p>
<ul>
<li>Fix: generation config empty for batch - <a href="https://github.com/BerriAI/litellm/pull/19556" target="_blank" rel="noopener noreferrer">PR #19556</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/pass_through">Pass Through Endpoints</a></strong></p>
<ul>
<li>Always reupdate registry - <a href="https://github.com/BerriAI/litellm/pull/19420" target="_blank" rel="noopener noreferrer">PR #19420</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-81-3#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-81-3#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h3>
<ul>
<li>
<p><strong>Cost Estimator</strong></p>
<ul>
<li>Fix model dropdown - <a href="https://github.com/BerriAI/litellm/pull/19529" target="_blank" rel="noopener noreferrer">PR #19529</a></li>
</ul>
</li>
<li>
<p><strong>Claude Code Plugins</strong></p>
<ul>
<li>Allow Adding Claude Code Plugins via UI - <a href="https://github.com/BerriAI/litellm/pull/19387" target="_blank" rel="noopener noreferrer">PR #19387</a></li>
</ul>
</li>
<li>
<p><strong>Guardrails</strong></p>
<ul>
<li>New Policy management UI - <a href="https://github.com/BerriAI/litellm/pull/19668" target="_blank" rel="noopener noreferrer">PR #19668</a></li>
<li>Allow adding policies on Keys/Teams + Viewing on Info panels - <a href="https://github.com/BerriAI/litellm/pull/19688" target="_blank" rel="noopener noreferrer">PR #19688</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>respects custom authentication header override - <a href="https://github.com/BerriAI/litellm/pull/19276" target="_blank" rel="noopener noreferrer">PR #19276</a></li>
</ul>
</li>
<li>
<p><strong>Playground</strong></p>
<ul>
<li>Button to Fill Custom API Base - <a href="https://github.com/BerriAI/litellm/pull/19440" target="_blank" rel="noopener noreferrer">PR #19440</a></li>
<li>display mcp output on the play ground - <a href="https://github.com/BerriAI/litellm/pull/19553" target="_blank" rel="noopener noreferrer">PR #19553</a></li>
</ul>
</li>
<li>
<p><strong>Models</strong></p>
<ul>
<li>Paginate /v2/models/info - <a href="https://github.com/BerriAI/litellm/pull/19521" target="_blank" rel="noopener noreferrer">PR #19521</a></li>
<li>All Model Tab Pagination - <a href="https://github.com/BerriAI/litellm/pull/19525" target="_blank" rel="noopener noreferrer">PR #19525</a></li>
<li>Adding Optional scope Param to /models - <a href="https://github.com/BerriAI/litellm/pull/19539" target="_blank" rel="noopener noreferrer">PR #19539</a></li>
<li>Model Search - <a href="https://github.com/BerriAI/litellm/pull/19622" target="_blank" rel="noopener noreferrer">PR #19622</a></li>
<li>Filter by Model ID and Team ID - <a href="https://github.com/BerriAI/litellm/pull/19713" target="_blank" rel="noopener noreferrer">PR #19713</a></li>
</ul>
</li>
<li>
<p><strong>MCP Servers</strong></p>
<ul>
<li>MCP Tools Tab Resetting to Overview - <a href="https://github.com/BerriAI/litellm/pull/19468" target="_blank" rel="noopener noreferrer">PR #19468</a></li>
</ul>
</li>
<li>
<p><strong>Organizations</strong></p>
<ul>
<li>Prevent org admin from creating a new user with proxy_admin permissions - <a href="https://github.com/BerriAI/litellm/pull/19296" target="_blank" rel="noopener noreferrer">PR #19296</a></li>
<li>Edit Page: Reusable Model Select - <a href="https://github.com/BerriAI/litellm/pull/19601" target="_blank" rel="noopener noreferrer">PR #19601</a></li>
</ul>
</li>
<li>
<p><strong>Teams</strong></p>
<ul>
<li>Reusable Model Select - <a href="https://github.com/BerriAI/litellm/pull/19543" target="_blank" rel="noopener noreferrer">PR #19543</a></li>
<li>[Fix] Team Update with Organization having All Proxy Models - <a href="https://github.com/BerriAI/litellm/pull/19604" target="_blank" rel="noopener noreferrer">PR #19604</a></li>
</ul>
</li>
<li>
<p><strong>Logs</strong></p>
<ul>
<li>Include tool arguments in spend logs table - <a href="https://github.com/BerriAI/litellm/pull/19640" target="_blank" rel="noopener noreferrer">PR #19640</a></li>
</ul>
</li>
<li>
<p><strong>Fallbacks / Loadbalancing</strong></p>
<ul>
<li>New fallbacks modal - <a href="https://github.com/BerriAI/litellm/pull/19673" target="_blank" rel="noopener noreferrer">PR #19673</a></li>
<li>Set fallbacks/loadbalancing by team/key - <a href="https://github.com/BerriAI/litellm/pull/19686" target="_blank" rel="noopener noreferrer">PR #19686</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bugs-1">Bugs<a href="https://docs.litellm.ai/release_notes/v1-81-3#bugs-1" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h3>
<ul>
<li>
<p><strong>Playground</strong></p>
<ul>
<li>increase model selector width in playground Compare view - <a href="https://github.com/BerriAI/litellm/pull/19423" target="_blank" rel="noopener noreferrer">PR #19423</a></li>
</ul>
</li>
<li>
<p><strong>Virtual Keys</strong></p>
<ul>
<li>Sorting Shows Incorrect Entries - <a href="https://github.com/BerriAI/litellm/pull/19534" target="_blank" rel="noopener noreferrer">PR #19534</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>UI 404 error when SERVER_ROOT_PATH is set - <a href="https://github.com/BerriAI/litellm/pull/19467" target="_blank" rel="noopener noreferrer">PR #19467</a></li>
<li>Redirect to ui/login on expired JWT - <a href="https://github.com/BerriAI/litellm/pull/19687" target="_blank" rel="noopener noreferrer">PR #19687</a></li>
</ul>
</li>
<li>
<p><strong>SSO</strong></p>
<ul>
<li>Fix SSO user roles not updating for existing users - <a href="https://github.com/BerriAI/litellm/pull/19621" target="_blank" rel="noopener noreferrer">PR #19621</a></li>
</ul>
</li>
<li>
<p><strong>Guardrails</strong></p>
<ul>
<li>ensure guardrail patterns persist on edit and mode toggle - <a href="https://github.com/BerriAI/litellm/pull/19265" target="_blank" rel="noopener noreferrer">PR #19265</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="ai-integrations">AI Integrations<a href="https://docs.litellm.ai/release_notes/v1-81-3#ai-integrations" class="hash-link" aria-label="Direct link to AI Integrations" title="Direct link to AI Integrations">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="logging">Logging<a href="https://docs.litellm.ai/release_notes/v1-81-3#logging" class="hash-link" aria-label="Direct link to Logging" title="Direct link to Logging">​</a></h3>
<ul>
<li><strong>General Logging</strong>
<ul>
<li>prevent printing duplicate StandardLoggingPayload logs - <a href="https://github.com/BerriAI/litellm/pull/19325" target="_blank" rel="noopener noreferrer">PR #19325</a></li>
<li>Fix: log duplication when json_logs is enabled - <a href="https://github.com/BerriAI/litellm/pull/19705" target="_blank" rel="noopener noreferrer">PR #19705</a></li>
</ul>
</li>
<li><strong>Langfuse OTEL</strong>
<ul>
<li>ignore service logs and fix callback shadowing - <a href="https://github.com/BerriAI/litellm/pull/19298" target="_blank" rel="noopener noreferrer">PR #19298</a></li>
</ul>
</li>
<li><strong>Langfuse</strong>
<ul>
<li>Send litellm_trace_id - <a href="https://github.com/BerriAI/litellm/pull/19528" target="_blank" rel="noopener noreferrer">PR #19528</a></li>
<li>Add Langfuse mock mode for testing without API calls - <a href="https://github.com/BerriAI/litellm/pull/19676" target="_blank" rel="noopener noreferrer">PR #19676</a></li>
</ul>
</li>
<li><strong>GCS Bucket</strong>
<ul>
<li>prevent unbounded queue growth due to slow API calls - <a href="https://github.com/BerriAI/litellm/pull/19297" target="_blank" rel="noopener noreferrer">PR #19297</a></li>
<li>Add GCS mock mode for testing without API calls - <a href="https://github.com/BerriAI/litellm/pull/19683" target="_blank" rel="noopener noreferrer">PR #19683</a></li>
</ul>
</li>
<li><strong>Responses API Logging</strong>
<ul>
<li>Fix pydantic serialization error - <a href="https://github.com/BerriAI/litellm/pull/19486" target="_blank" rel="noopener noreferrer">PR #19486</a></li>
</ul>
</li>
<li><strong>Arize Phoenix</strong>
<ul>
<li>add openinference span kinds to arize phoenix - <a href="https://github.com/BerriAI/litellm/pull/19267" target="_blank" rel="noopener noreferrer">PR #19267</a></li>
</ul>
</li>
<li><strong>Prometheus</strong>
<ul>
<li>Added new prometheus metrics for user count and team count - <a href="https://github.com/BerriAI/litellm/pull/19520" target="_blank" rel="noopener noreferrer">PR #19520</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails">Guardrails<a href="https://docs.litellm.ai/release_notes/v1-81-3#guardrails" class="hash-link" aria-label="Direct link to Guardrails" title="Direct link to Guardrails">​</a></h3>
<ul>
<li><strong>Bedrock Guardrails</strong>
<ul>
<li>Ensure post_call guardrail checks input+output - <a href="https://github.com/BerriAI/litellm/pull/19151" target="_blank" rel="noopener noreferrer">PR #19151</a></li>
</ul>
</li>
<li><strong>Prompt Security</strong>
<ul>
<li>fixing prompt-security's guardrail implementation - <a href="https://github.com/BerriAI/litellm/pull/19374" target="_blank" rel="noopener noreferrer">PR #19374</a></li>
</ul>
</li>
<li><strong>Presidio</strong>
<ul>
<li>Fixes crash in Presidio Guardrail when running in background threads (logging_hook) - <a href="https://github.com/BerriAI/litellm/pull/19714" target="_blank" rel="noopener noreferrer">PR #19714</a></li>
</ul>
</li>
<li><strong>Pillar Security</strong>
<ul>
<li>Migrate Pillar Security to Generic Guardrail API - <a href="https://github.com/BerriAI/litellm/pull/19364" target="_blank" rel="noopener noreferrer">PR #19364</a></li>
</ul>
</li>
<li><strong>Policy Engine</strong>
<ul>
<li>New LiteLLM Policy engine - create policies to manage guardrails, conditions - permissions per Key, Team - <a href="https://github.com/BerriAI/litellm/pull/19612" target="_blank" rel="noopener noreferrer">PR #19612</a></li>
</ul>
</li>
<li><strong>General</strong>
<ul>
<li>add case-insensitive support for guardrail mode and actions - <a href="https://github.com/BerriAI/litellm/pull/19480" target="_blank" rel="noopener noreferrer">PR #19480</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="prompt-management">Prompt Management<a href="https://docs.litellm.ai/release_notes/v1-81-3#prompt-management" class="hash-link" aria-label="Direct link to Prompt Management" title="Direct link to Prompt Management">​</a></h3>
<ul>
<li><strong>General</strong>
<ul>
<li>fix prompt info lookup and delete using correct IDs - <a href="https://github.com/BerriAI/litellm/pull/19358" target="_blank" rel="noopener noreferrer">PR #19358</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="secret-manager">Secret Manager<a href="https://docs.litellm.ai/release_notes/v1-81-3#secret-manager" class="hash-link" aria-label="Direct link to Secret Manager" title="Direct link to Secret Manager">​</a></h3>
<ul>
<li><strong>AWS Secret Manager</strong>
<ul>
<li>ensure auto-rotation updates existing AWS secret instead of creating new one - <a href="https://github.com/BerriAI/litellm/pull/19455" target="_blank" rel="noopener noreferrer">PR #19455</a></li>
</ul>
</li>
<li><strong>Hashicorp Vault</strong>
<ul>
<li>Ensure key rotations work with Vault - <a href="https://github.com/BerriAI/litellm/pull/19634" target="_blank" rel="noopener noreferrer">PR #19634</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-81-3#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li><strong>Pricing Updates</strong>
<ul>
<li>Add openai/dall-e base pricing entries - <a href="https://github.com/BerriAI/litellm/pull/19133" target="_blank" rel="noopener noreferrer">PR #19133</a></li>
<li>Add <code>input_cost_per_video_per_second</code> in ModelInfoBase - <a href="https://github.com/BerriAI/litellm/pull/19398" target="_blank" rel="noopener noreferrer">PR #19398</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements">Performance / Loadbalancing / Reliability improvements<a href="https://docs.litellm.ai/release_notes/v1-81-3#performance--loadbalancing--reliability-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements" title="Direct link to Performance / Loadbalancing / Reliability improvements">​</a></h2>
<ul>
<li>
<p><strong>General</strong></p>
<ul>
<li>Fix date overflow/division by zero in proxy utils - <a href="https://github.com/BerriAI/litellm/pull/19527" target="_blank" rel="noopener noreferrer">PR #19527</a></li>
<li>Fix in-flight request termination on SIGTERM when health-check runs in a separate process - <a href="https://github.com/BerriAI/litellm/pull/19427" target="_blank" rel="noopener noreferrer">PR #19427</a></li>
<li>Fix Pass through routes to work with server root path - <a href="https://github.com/BerriAI/litellm/pull/19383" target="_blank" rel="noopener noreferrer">PR #19383</a></li>
<li>Fix logging error for stop iteration - <a href="https://github.com/BerriAI/litellm/pull/19649" target="_blank" rel="noopener noreferrer">PR #19649</a></li>
<li>prevent retrying 4xx client errors - <a href="https://github.com/BerriAI/litellm/pull/19275" target="_blank" rel="noopener noreferrer">PR #19275</a></li>
<li>add better error handling for misconfig on health check - <a href="https://github.com/BerriAI/litellm/pull/19441" target="_blank" rel="noopener noreferrer">PR #19441</a></li>
</ul>
</li>
<li>
<p><strong>Router</strong></p>
<ul>
<li>Fix Azure RPM calculation formula - <a href="https://github.com/BerriAI/litellm/pull/19513" target="_blank" rel="noopener noreferrer">PR #19513</a></li>
<li>Persist scheduler request queue to redis - <a href="https://github.com/BerriAI/litellm/pull/19304" target="_blank" rel="noopener noreferrer">PR #19304</a></li>
<li>pass search_tools to Router during DB-triggered initialization - <a href="https://github.com/BerriAI/litellm/pull/19388" target="_blank" rel="noopener noreferrer">PR #19388</a></li>
<li>Fixed PromptCachingCache to correctly handle messages where cache_control is a sibling key of string content - <a href="https://github.com/BerriAI/litellm/pull/19266" target="_blank" rel="noopener noreferrer">PR #19266</a></li>
</ul>
</li>
<li>
<p><strong>Memory Leaks/OOM</strong></p>
<ul>
<li>prevent OOM with nested $defs in tool schemas - <a href="https://github.com/BerriAI/litellm/pull/19112" target="_blank" rel="noopener noreferrer">PR #19112</a></li>
<li>fix: HTTP client memory leaks in Presidio, OpenAI, and Gemini - <a href="https://github.com/BerriAI/litellm/pull/19190" target="_blank" rel="noopener noreferrer">PR #19190</a></li>
</ul>
</li>
<li>
<p><strong>Non root</strong></p>
<ul>
<li>fix logfile and pidfile of supervisor for non root environment - <a href="https://github.com/BerriAI/litellm/pull/17267" target="_blank" rel="noopener noreferrer">PR #17267</a></li>
<li>resolve Read-only file system error in non-root images - <a href="https://github.com/BerriAI/litellm/pull/19449" target="_blank" rel="noopener noreferrer">PR #19449</a></li>
</ul>
</li>
<li>
<p><strong>Dockerfile</strong></p>
<ul>
<li>Redis Semantic Caching - add missing redisvl dependency to requirements.txt - <a href="https://github.com/BerriAI/litellm/pull/19417" target="_blank" rel="noopener noreferrer">PR #19417</a></li>
<li>Bump OTEL versions to support a2a dependency - resolves modulenotfounderror for Microsoft Agents by&nbsp;@Harshit28j&nbsp;in&nbsp;#18991</li>
</ul>
</li>
<li>
<p><strong>DB</strong></p>
<ul>
<li>Handle PostgreSQL cached plan errors during rolling deployments - <a href="https://github.com/BerriAI/litellm/pull/19424" target="_blank" rel="noopener noreferrer">PR #19424</a></li>
</ul>
</li>
<li>
<p><strong>Timeouts</strong></p>
<ul>
<li>Fix: total timeout is not respected - <a href="https://github.com/BerriAI/litellm/pull/19389" target="_blank" rel="noopener noreferrer">PR #19389</a></li>
</ul>
</li>
<li>
<p><strong>SDK</strong></p>
<ul>
<li>Field-Existence Checks to Type Classes to Prevent Attribute Errors - <a href="https://github.com/BerriAI/litellm/pull/18321" target="_blank" rel="noopener noreferrer">PR #18321</a></li>
<li>add google-cloud-aiplatform as optional dependency with clear error message - <a href="https://github.com/BerriAI/litellm/pull/19437" target="_blank" rel="noopener noreferrer">PR #19437</a></li>
<li>Make&nbsp;grpc&nbsp;dependency optional - <a href="https://github.com/BerriAI/litellm/pull/19447" target="_blank" rel="noopener noreferrer">PR #19447</a></li>
<li>Add support for retry policies - <a href="https://github.com/BerriAI/litellm/pull/19645" target="_blank" rel="noopener noreferrer">PR #19645</a></li>
</ul>
</li>
<li>
<p><strong>Performance</strong></p>
<ul>
<li>Cut chat_completion latency by ~21% by reducing pre-call processing time - <a href="https://github.com/BerriAI/litellm/pull/19535" target="_blank" rel="noopener noreferrer">PR #19535</a></li>
<li>Optimize strip_trailing_slash with O(1) index check - <a href="https://github.com/BerriAI/litellm/pull/19679" target="_blank" rel="noopener noreferrer">PR #19679</a></li>
<li>Optimize use_custom_pricing_for_model with set intersection - <a href="https://github.com/BerriAI/litellm/pull/19677" target="_blank" rel="noopener noreferrer">PR #19677</a></li>
<li>perf: skip pattern_router.route() for non-wildcard models - <a href="https://github.com/BerriAI/litellm/pull/19664" target="_blank" rel="noopener noreferrer">PR #19664</a></li>
<li>perf: Add LRU caching to get_model_info for faster cost lookups - <a href="https://github.com/BerriAI/litellm/pull/19606" target="_blank" rel="noopener noreferrer">PR #19606</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="general-proxy-improvements">General Proxy Improvements<a href="https://docs.litellm.ai/release_notes/v1-81-3#general-proxy-improvements" class="hash-link" aria-label="Direct link to General Proxy Improvements" title="Direct link to General Proxy Improvements">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="doc-improvements">Doc Improvements<a href="https://docs.litellm.ai/release_notes/v1-81-3#doc-improvements" class="hash-link" aria-label="Direct link to Doc Improvements" title="Direct link to Doc Improvements">​</a></h3>
<ul>
<li>new tutorial for adding MCPs to Cursor via LiteLLM - <a href="https://github.com/BerriAI/litellm/pull/19317" target="_blank" rel="noopener noreferrer">PR #19317</a></li>
<li>fix vertex_region to vertex_location in Vertex AI pass-through docs - <a href="https://github.com/BerriAI/litellm/pull/19380" target="_blank" rel="noopener noreferrer">PR #19380</a></li>
<li>clarify Gemini and Vertex AI model prefix in json file - <a href="https://github.com/BerriAI/litellm/pull/19443" target="_blank" rel="noopener noreferrer">PR #19443</a></li>
<li>update Claude Code integration guides - <a href="https://github.com/BerriAI/litellm/pull/19415" target="_blank" rel="noopener noreferrer">PR #19415</a></li>
<li>adjust opencode tutorial - <a href="https://github.com/BerriAI/litellm/pull/19605" target="_blank" rel="noopener noreferrer">PR #19605</a></li>
<li>add spend-queue-troubleshooting docs - <a href="https://github.com/BerriAI/litellm/pull/19659" target="_blank" rel="noopener noreferrer">PR #19659</a></li>
<li>docs: add litellm-enterprise requirement for managed files - <a href="https://github.com/BerriAI/litellm/pull/19689" target="_blank" rel="noopener noreferrer">PR #19689</a></li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="helm">Helm<a href="https://docs.litellm.ai/release_notes/v1-81-3#helm" class="hash-link" aria-label="Direct link to Helm" title="Direct link to Helm">​</a></h3>
<ul>
<li>Add support for keda in helm chart - <a href="https://github.com/BerriAI/litellm/pull/19337" target="_blank" rel="noopener noreferrer">PR #19337</a></li>
<li>sync Helm chart version with LiteLLM release version - <a href="https://github.com/BerriAI/litellm/pull/19438" target="_blank" rel="noopener noreferrer">PR #19438</a></li>
<li>Enable PreStop hook configuration in values.yaml - <a href="https://github.com/BerriAI/litellm/pull/19613" target="_blank" rel="noopener noreferrer">PR #19613</a></li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="general">General<a href="https://docs.litellm.ai/release_notes/v1-81-3#general" class="hash-link" aria-label="Direct link to General" title="Direct link to General">​</a></h3>
<ul>
<li>Add health check scripts and parallel execution support - <a href="https://github.com/BerriAI/litellm/pull/19295" target="_blank" rel="noopener noreferrer">PR #19295</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-81-3#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@dushyantzz made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19158" target="_blank" rel="noopener noreferrer">PR #19158</a></li>
<li>@obod-mpw made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19133" target="_blank" rel="noopener noreferrer">PR #19133</a></li>
<li>@msexxeta made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19030" target="_blank" rel="noopener noreferrer">PR #19030</a></li>
<li>@rsicart made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19337" target="_blank" rel="noopener noreferrer">PR #19337</a></li>
<li>@cluebbehusen made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19311" target="_blank" rel="noopener noreferrer">PR #19311</a></li>
<li>@Lucky-Lodhi2004 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19315" target="_blank" rel="noopener noreferrer">PR #19315</a></li>
<li>@binbandit made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19324" target="_blank" rel="noopener noreferrer">PR #19324</a></li>
<li>@flex-myeonghyeon made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19381" target="_blank" rel="noopener noreferrer">PR #19381</a></li>
<li>@Lrakotoson made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18321" target="_blank" rel="noopener noreferrer">PR #18321</a></li>
<li>@bensi94 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18787" target="_blank" rel="noopener noreferrer">PR #18787</a></li>
<li>@victorigualada made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19368" target="_blank" rel="noopener noreferrer">PR #19368</a></li>
<li>@VedantMadane made their first contribution in #19266</li>
<li>@stiyyagura0901 made their first contribution in #19276</li>
<li>@kamilio made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19447" target="_blank" rel="noopener noreferrer">PR #19447</a></li>
<li>@jonathansampson made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19433" target="_blank" rel="noopener noreferrer">PR #19433</a></li>
<li>@rynecarbone made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19416" target="_blank" rel="noopener noreferrer">PR #19416</a></li>
<li>@jayy-77 made their first contribution in #19366</li>
<li>@davida-ps made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19374" target="_blank" rel="noopener noreferrer">PR #19374</a></li>
<li>@joaodinissf made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19506" target="_blank" rel="noopener noreferrer">PR #19506</a></li>
<li>@ecao310 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19520" target="_blank" rel="noopener noreferrer">PR #19520</a></li>
<li>@mpcusack-altos made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19577" target="_blank" rel="noopener noreferrer">PR #19577</a></li>
<li>@milan-berri made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19602" target="_blank" rel="noopener noreferrer">PR #19602</a></li>
<li>@xqe2011 made their first contribution in #19621</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog">Full Changelog<a href="https://docs.litellm.ai/release_notes/v1-81-3#full-changelog" class="hash-link" aria-label="Direct link to Full Changelog" title="Direct link to Full Changelog">​</a></h2>
<p><strong><a href="https://github.com/BerriAI/litellm/releases/tag/v1.81.3.rc" target="_blank" rel="noopener noreferrer">View complete changelog on GitHub</a></strong></p>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[v1.81.0-stable - Claude Code - Web Search Across All Providers]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-81-0</id>
        <link href="https://docs.litellm.ai/release_notes/v1-81-0"/>
        <updated>2026-01-18T10:00:00.000Z</updated>
        <summary type="html"><![CDATA[Deploy this version]]></summary>
        <content type="html"><![CDATA[<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-81-0#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">docker.litellm.ai/berriai/litellm:v1.81.0-stable</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.81.0</span><br></span></code></pre></div></div></div></div></div>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-81-0#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<ul>
<li><strong>Claude Code</strong> - Support for using web search across Bedrock, Vertex AI, and all LiteLLM providers</li>
<li><strong>Major Change</strong> - <a href="https://docs.litellm.ai/release_notes/v1-81-0#major-change---chatcompletions-image-url-download-size-limit">50MB limit on image URL downloads</a> to improve reliability</li>
<li><strong>Performance</strong> - <a href="https://docs.litellm.ai/release_notes/v1-81-0#performance---25-cpu-usage-reduction">25% CPU Usage Reduction</a> by removing premature model.dump() calls from the hot path</li>
<li><strong>Deleted Keys Audit Table on UI</strong> - <a href="https://docs.litellm.ai/docs/proxy/deleted_keys_teams.md">View deleted keys and teams for audit purposes</a> with spend and budget information at the time of deletion</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="claude-code---web-search-across-all-providers">Claude Code - Web Search Across All Providers<a href="https://docs.litellm.ai/release_notes/v1-81-0#claude-code---web-search-across-all-providers" class="hash-link" aria-label="Direct link to Claude Code - Web Search Across All Providers" title="Direct link to Claude Code - Web Search Across All Providers">​</a></h2>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAFCAYAAAB8ZH1oAAAACXBIWXMAACxLAAAsSwGlPZapAAAAnElEQVR4nDWKO07DQBQAn/f3doNXG8dfbCJEQ0VcJQUo10jKHITLD5IjitEUM/K4/bIMP8zjlZLPqPsihRX1p83/yGW9c+y/Cb6hyQt1HIi+palndnEgaU/UHon+RK2fOGfp9jVteUHVE4PFGsFa2Sw7XQn+g0PbMc1vdP3A9LpQSkGDwzuDNdVz9PZ9C9M4knMmpYQxhqqqEJGNP+RsO/JzMjigAAAAAElFTkSuQmCC&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="334"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/claude_code_websearch.9a9c1d7.640.png" srcset="/assets/ideal-img/claude_code_websearch.9a9c1d7.640.png 640w,/assets/ideal-img/claude_code_websearch.e852a7e.1920.png 1920w" width="640" height="334"></noscript></div>
<p>This release brings web search support to Claude Code across all LiteLLM providers (Bedrock, Azure, Vertex AI, and more), enabling AI coding assistants to search the web for real-time information.</p>
<p>This means you can now use Claude Code's web search tool with any provider, not just Anthropic's native API. LiteLLM automatically intercepts web search requests and executes them server-side using your configured search provider (Perplexity, Tavily, Exa AI, and more).</p>
<p>Proxy Admins can configure web search interception in their LiteLLM proxy config to enable this capability for their teams using Claude Code with Bedrock, Azure, or any other supported provider.</p>
<p><a href="https://docs.litellm.ai/docs/tutorials/claude_code_websearch" target="_blank" rel="noopener noreferrer"><strong>Learn more →</strong></a></p>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="major-change---chatcompletions-image-url-download-size-limit">Major Change - /chat/completions Image URL Download Size Limit<a href="https://docs.litellm.ai/release_notes/v1-81-0#major-change---chatcompletions-image-url-download-size-limit" class="hash-link" aria-label="Direct link to Major Change - /chat/completions Image URL Download Size Limit" title="Direct link to Major Change - /chat/completions Image URL Download Size Limit">​</a></h2>
<p>To improve reliability and prevent memory issues, LiteLLM now includes a configurable <strong>50MB limit</strong> on image URL downloads by default. Previously, there was no limit on image downloads, which could occasionally cause memory issues with very large images.</p>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="how-it-works">How It Works<a href="https://docs.litellm.ai/release_notes/v1-81-0#how-it-works" class="hash-link" aria-label="Direct link to How It Works" title="Direct link to How It Works">​</a></h3>
<p>Requests with image URLs exceeding 50MB will receive a helpful error message:</p>
<div class="language-bash codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-bash codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">curl -X POST 'https://your-litellm-proxy.com/chat/completions' \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">  -H 'Content-Type: application/json' \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">  -H 'Authorization: Bearer sk-1234' \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">  -d '{</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">    "model": "gpt-4o",</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">    "messages": [</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">      {</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">        "role": "user",</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">        "content": [</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">          {</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">            "type": "text",</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">            "text": "What is in this image?"</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">          },</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">          {</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">            "type": "image_url",</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">            "image_url": {</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">              "url": "https://example.com/very-large-image.jpg"</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">            }</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">          }</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">        ]</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">      }</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">    ]</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">  }'</span><br></span></code></pre></div></div>
<p><strong>Error Response:</strong></p>
<div class="language-json codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-json codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token punctuation" style="color:#393A34">{</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">  </span><span class="token property" style="color:#36acaa">"error"</span><span class="token operator" style="color:#393A34">:</span><span class="token plain"> </span><span class="token punctuation" style="color:#393A34">{</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">    </span><span class="token property" style="color:#36acaa">"message"</span><span class="token operator" style="color:#393A34">:</span><span class="token plain"> </span><span class="token string" style="color:#e3116c">"Error: Image size (75.50MB) exceeds maximum allowed size (50.0MB). url=https://example.com/very-large-image.jpg"</span><span class="token punctuation" style="color:#393A34">,</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">    </span><span class="token property" style="color:#36acaa">"type"</span><span class="token operator" style="color:#393A34">:</span><span class="token plain"> </span><span class="token string" style="color:#e3116c">"ImageFetchError"</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">  </span><span class="token punctuation" style="color:#393A34">}</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain"></span><span class="token punctuation" style="color:#393A34">}</span><br></span></code></pre></div></div>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="configuring-the-limit">Configuring the Limit<a href="https://docs.litellm.ai/release_notes/v1-81-0#configuring-the-limit" class="hash-link" aria-label="Direct link to Configuring the Limit" title="Direct link to Configuring the Limit">​</a></h3>
<p>The default 50MB limit works well for most use cases, but you can easily adjust it if needed:</p>
<p><strong>Increase the limit (e.g., to 100MB):</strong></p>
<div class="language-bash codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-bash codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">export MAX_IMAGE_URL_DOWNLOAD_SIZE_MB=100</span><br></span></code></pre></div></div>
<p><strong>Disable image URL downloads (for security):</strong></p>
<div class="language-bash codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-bash codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">export MAX_IMAGE_URL_DOWNLOAD_SIZE_MB=0</span><br></span></code></pre></div></div>
<p><strong>Docker Configuration:</strong></p>
<div class="language-bash codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-bash codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">  -e MAX_IMAGE_URL_DOWNLOAD_SIZE_MB=100 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">  -p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">  docker.litellm.ai/berriai/litellm:v1.81.0</span><br></span></code></pre></div></div>
<p><strong>Proxy Config (config.yaml):</strong></p>
<div class="language-yaml codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-yaml codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token key atrule" style="color:#00a4db">general_settings</span><span class="token punctuation" style="color:#393A34">:</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">  </span><span class="token key atrule" style="color:#00a4db">master_key</span><span class="token punctuation" style="color:#393A34">:</span><span class="token plain"> sk</span><span class="token punctuation" style="color:#393A34">-</span><span class="token number" style="color:#36acaa">1234</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">  </span><br></span><span class="token-line" style="color:#393A34"><span class="token plain"></span><span class="token comment" style="color:#999988;font-style:italic"># Set via environment variable</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain"></span><span class="token key atrule" style="color:#00a4db">environment_variables</span><span class="token punctuation" style="color:#393A34">:</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">  </span><span class="token key atrule" style="color:#00a4db">MAX_IMAGE_URL_DOWNLOAD_SIZE_MB</span><span class="token punctuation" style="color:#393A34">:</span><span class="token plain"> </span><span class="token string" style="color:#e3116c">"100"</span><br></span></code></pre></div></div>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="why-add-this">Why Add This?<a href="https://docs.litellm.ai/release_notes/v1-81-0#why-add-this" class="hash-link" aria-label="Direct link to Why Add This?" title="Direct link to Why Add This?">​</a></h3>
<p>This feature improves reliability by:</p>
<ul>
<li>Preventing memory issues from very large images</li>
<li>Aligning with OpenAI's 50MB payload limit</li>
<li>Validating image sizes early (when Content-Length header is available)</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance---25-cpu-usage-reduction">Performance - 25% CPU Usage Reduction<a href="https://docs.litellm.ai/release_notes/v1-81-0#performance---25-cpu-usage-reduction" class="hash-link" aria-label="Direct link to Performance - 25% CPU Usage Reduction" title="Direct link to Performance - 25% CPU Usage Reduction">​</a></h2>
<p>LiteLLM now reduces CPU usage by removing premature <code>model.dump()</code> calls from the hot path in request processing. Previously, Pydantic model serialization was performed earlier and more frequently than necessary, causing unnecessary CPU overhead on every request. By deferring serialization until it is actually needed, LiteLLM reduces CPU usage and improves request throughput under high load.</p>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deleted-keys-audit-table-on-ui">Deleted Keys Audit Table on UI<a href="https://docs.litellm.ai/release_notes/v1-81-0#deleted-keys-audit-table-on-ui" class="hash-link" aria-label="Direct link to Deleted Keys Audit Table on UI" title="Direct link to Deleted Keys Audit Table on UI">​</a></h2>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAFCAYAAAB8ZH1oAAAACXBIWXMAACxLAAAsSwGlPZapAAAAiUlEQVR4nC3NSw6DMBAEUe5/Q4QE2YZPhI1hbM8MFTmh10/V3TAM9H3PNE2M48hrGpHrxExRLZRSqKXQxRhY1415WYghMG+ZeVfSVfgkB27cbzozQ0QQyagaOSu5KGbOeT3FWulUK0mUd3C2A9IpHMeBu3N7w/ILPNB4784SbtJVifEP29pTq38BrvHBNSH/Hh8AAAAASUVORK5CYII=&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="334"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/ui_deleted_keys_table.98dcbba.640.png" srcset="/assets/ideal-img/ui_deleted_keys_table.98dcbba.640.png 640w,/assets/ideal-img/ui_deleted_keys_table.463df6b.1920.png 1920w" width="640" height="334"></noscript></div>
<p>LiteLLM now provides a comprehensive audit table for deleted API keys and teams directly in the UI. This feature allows you to easily track the spend of deleted keys, view their associated team information, and maintain accurate financial records for auditing and compliance purposes. The table displays key details including key aliases, team associations, and spend information captured at the time of deletion. For more information on how to use this feature, see the <a href="https://docs.litellm.ai/docs/proxy/deleted_keys_teams.md">Deleted Keys &amp; Teams documentation</a>.</p>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-81-0#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support">New Model Support<a href="https://docs.litellm.ai/release_notes/v1-81-0#new-model-support" class="hash-link" aria-label="Direct link to New Model Support" title="Direct link to New Model Support">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Features</th></tr></thead><tbody><tr><td>OpenAI</td><td><code>gpt-5.2-codex</code></td><td>Code generation</td></tr><tr><td>Azure</td><td><code>azure/gpt-5.2-codex</code></td><td>Code generation</td></tr><tr><td>Cerebras</td><td><code>cerebras/zai-glm-4.7</code></td><td>Reasoning, function calling</td></tr><tr><td>Replicate</td><td>All chat models</td><td>Full support for all Replicate chat models</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-81-0#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Add missing anthropic tool results in response - <a href="https://github.com/BerriAI/litellm/pull/18945" target="_blank" rel="noopener noreferrer">PR #18945</a></li>
<li>Preserve web_fetch_tool_result in multi-turn conversations - <a href="https://github.com/BerriAI/litellm/pull/18142" target="_blank" rel="noopener noreferrer">PR #18142</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Gemini</a></strong></p>
<ul>
<li>Add presence_penalty support for Google AI Studio - <a href="https://github.com/BerriAI/litellm/pull/18154" target="_blank" rel="noopener noreferrer">PR #18154</a></li>
<li>Forward extra_headers in generateContent adapter - <a href="https://github.com/BerriAI/litellm/pull/18935" target="_blank" rel="noopener noreferrer">PR #18935</a></li>
<li>Add medium value support for detail param - <a href="https://github.com/BerriAI/litellm/pull/19187" target="_blank" rel="noopener noreferrer">PR #19187</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vertex">Vertex AI</a></strong></p>
<ul>
<li>Improve passthrough endpoint URL parsing and construction - <a href="https://github.com/BerriAI/litellm/pull/17526" target="_blank" rel="noopener noreferrer">PR #17526</a></li>
<li>Add type object to tool schemas missing type field - <a href="https://github.com/BerriAI/litellm/pull/19103" target="_blank" rel="noopener noreferrer">PR #19103</a></li>
<li>Keep type field in Gemini schema when properties is empty - <a href="https://github.com/BerriAI/litellm/pull/18979" target="_blank" rel="noopener noreferrer">PR #18979</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong></p>
<ul>
<li>Add OpenAI-compatible service_tier parameter translation - <a href="https://github.com/BerriAI/litellm/pull/18091" target="_blank" rel="noopener noreferrer">PR #18091</a></li>
<li>Add user auth in standard logging object for Bedrock passthrough - <a href="https://github.com/BerriAI/litellm/pull/19140" target="_blank" rel="noopener noreferrer">PR #19140</a></li>
<li>Strip throughput tier suffixes from model names - <a href="https://github.com/BerriAI/litellm/pull/19147" target="_blank" rel="noopener noreferrer">PR #19147</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/oci">OCI</a></strong></p>
<ul>
<li>Handle OpenAI-style image_url object in multimodal messages - <a href="https://github.com/BerriAI/litellm/pull/18272" target="_blank" rel="noopener noreferrer">PR #18272</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/ollama">Ollama</a></strong></p>
<ul>
<li>Set finish_reason to tool_calls and remove broken capability check - <a href="https://github.com/BerriAI/litellm/pull/18924" target="_blank" rel="noopener noreferrer">PR #18924</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/watsonx/index">Watsonx</a></strong></p>
<ul>
<li>Allow passing scope ID for Watsonx inferencing - <a href="https://github.com/BerriAI/litellm/pull/18959" target="_blank" rel="noopener noreferrer">PR #18959</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/replicate">Replicate</a></strong></p>
<ul>
<li>Add all chat Replicate models support - <a href="https://github.com/BerriAI/litellm/pull/18954" target="_blank" rel="noopener noreferrer">PR #18954</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openrouter">OpenRouter</a></strong></p>
<ul>
<li>Add OpenRouter support for image/generation endpoints - <a href="https://github.com/BerriAI/litellm/pull/19059" target="_blank" rel="noopener noreferrer">PR #19059</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/volcano">Volcengine</a></strong></p>
<ul>
<li>Add max_tokens settings for Volcengine models (deepseek-v3-2, glm-4-7, kimi-k2-thinking) - <a href="https://github.com/BerriAI/litellm/pull/19076" target="_blank" rel="noopener noreferrer">PR #19076</a></li>
</ul>
</li>
<li>
<p><strong>Azure Model Router</strong></p>
<ul>
<li>New Model - Azure Model Router on LiteLLM AI Gateway - <a href="https://github.com/BerriAI/litellm/pull/19054" target="_blank" rel="noopener noreferrer">PR #19054</a></li>
</ul>
</li>
<li>
<p><strong>GPT-5 Models</strong></p>
<ul>
<li>Correct context window sizes for GPT-5 model variants - <a href="https://github.com/BerriAI/litellm/pull/18928" target="_blank" rel="noopener noreferrer">PR #18928</a></li>
<li>Correct max_input_tokens for GPT-5 models - <a href="https://github.com/BerriAI/litellm/pull/19056" target="_blank" rel="noopener noreferrer">PR #19056</a></li>
</ul>
</li>
<li>
<p><strong>Text Completion</strong></p>
<ul>
<li>Support token IDs (list of integers) as prompt - <a href="https://github.com/BerriAI/litellm/pull/18011" target="_blank" rel="noopener noreferrer">PR #18011</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-81-0#bug-fixes" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Prevent dropping thinking when any message has thinking_blocks - <a href="https://github.com/BerriAI/litellm/pull/18929" target="_blank" rel="noopener noreferrer">PR #18929</a></li>
<li>Fix anthropic token counter with thinking - <a href="https://github.com/BerriAI/litellm/pull/19067" target="_blank" rel="noopener noreferrer">PR #19067</a></li>
<li>Add better error handling for Anthropic - <a href="https://github.com/BerriAI/litellm/pull/18955" target="_blank" rel="noopener noreferrer">PR #18955</a></li>
<li>Fix Anthropic during call error - <a href="https://github.com/BerriAI/litellm/pull/19060" target="_blank" rel="noopener noreferrer">PR #19060</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Gemini</a></strong></p>
<ul>
<li>Fix missing <code>completion_tokens_details</code> in Gemini 3 Flash when reasoning_effort is not used - <a href="https://github.com/BerriAI/litellm/pull/18898" target="_blank" rel="noopener noreferrer">PR #18898</a></li>
<li>Fix Gemini Image Generation imageConfig parameters - <a href="https://github.com/BerriAI/litellm/pull/18948" target="_blank" rel="noopener noreferrer">PR #18948</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vertex">Vertex AI</a></strong></p>
<ul>
<li>Fix Vertex AI 400 Error with CachedContent model mismatch - <a href="https://github.com/BerriAI/litellm/pull/19193" target="_blank" rel="noopener noreferrer">PR #19193</a></li>
<li>Fix Vertex AI doesn't support structured output - <a href="https://github.com/BerriAI/litellm/pull/19201" target="_blank" rel="noopener noreferrer">PR #19201</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong></p>
<ul>
<li>Fix Claude Code (<code>/messages</code>) Bedrock Invoke usage and request signing - <a href="https://github.com/BerriAI/litellm/pull/19111" target="_blank" rel="noopener noreferrer">PR #19111</a></li>
<li>Fix model ID encoding for Bedrock passthrough - <a href="https://github.com/BerriAI/litellm/pull/18944" target="_blank" rel="noopener noreferrer">PR #18944</a></li>
<li>Respect max_completion_tokens in thinking feature - <a href="https://github.com/BerriAI/litellm/pull/18946" target="_blank" rel="noopener noreferrer">PR #18946</a></li>
<li>Fix header forwarding in Bedrock passthrough - <a href="https://github.com/BerriAI/litellm/pull/19007" target="_blank" rel="noopener noreferrer">PR #19007</a></li>
<li>Fix Bedrock stability model usage issues - <a href="https://github.com/BerriAI/litellm/pull/19199" target="_blank" rel="noopener noreferrer">PR #19199</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-81-0#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-81-0#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">/messages (Claude Code)</a></strong></p>
<ul>
<li>Add support for Tool Search on <code>/messages</code> API across Azure, Bedrock, and Anthropic API - <a href="https://github.com/BerriAI/litellm/pull/19165" target="_blank" rel="noopener noreferrer">PR #19165</a></li>
<li>Track end-users with Claude Code (<code>/messages</code>) for better analytics and monitoring - <a href="https://github.com/BerriAI/litellm/pull/19171" target="_blank" rel="noopener noreferrer">PR #19171</a></li>
<li>Add web search support using LiteLLM <code>/search</code> endpoint with Claude Code (<code>/messages</code>) - <a href="https://github.com/BerriAI/litellm/pull/19263" target="_blank" rel="noopener noreferrer">PR #19263</a>, <a href="https://github.com/BerriAI/litellm/pull/19294" target="_blank" rel="noopener noreferrer">PR #19294</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">/messages (Claude Code) - Bedrock</a></strong></p>
<ul>
<li>Add support for Prompt Caching with Bedrock Converse on <code>/messages</code> - <a href="https://github.com/BerriAI/litellm/pull/19123" target="_blank" rel="noopener noreferrer">PR #19123</a></li>
<li>Ensure budget tokens are passed to Bedrock Converse API correctly on <code>/messages</code> - <a href="https://github.com/BerriAI/litellm/pull/19107" target="_blank" rel="noopener noreferrer">PR #19107</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/response_api">Responses API</a></strong></p>
<ul>
<li>Add support for caching for responses API - <a href="https://github.com/BerriAI/litellm/pull/19068" target="_blank" rel="noopener noreferrer">PR #19068</a></li>
<li>Add retry policy support to responses API - <a href="https://github.com/BerriAI/litellm/pull/19074" target="_blank" rel="noopener noreferrer">PR #19074</a></li>
</ul>
</li>
<li>
<p><strong>Realtime API</strong></p>
<ul>
<li>Use non-streaming method for endpoint v1/a2a/message/send - <a href="https://github.com/BerriAI/litellm/pull/19025" target="_blank" rel="noopener noreferrer">PR #19025</a></li>
</ul>
</li>
<li>
<p><strong>Batch API</strong></p>
<ul>
<li>Fix batch deletion and retrieve - <a href="https://github.com/BerriAI/litellm/pull/18340" target="_blank" rel="noopener noreferrer">PR #18340</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs">Bugs<a href="https://docs.litellm.ai/release_notes/v1-81-0#bugs" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>General</strong>
<ul>
<li>Fix responses content can't be none - <a href="https://github.com/BerriAI/litellm/pull/19064" target="_blank" rel="noopener noreferrer">PR #19064</a></li>
<li>Fix model name from query param in realtime request - <a href="https://github.com/BerriAI/litellm/pull/19135" target="_blank" rel="noopener noreferrer">PR #19135</a></li>
<li>Fix video status/content credential injection for wildcard models - <a href="https://github.com/BerriAI/litellm/pull/18854" target="_blank" rel="noopener noreferrer">PR #18854</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-81-0#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-81-0#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<p><strong>Virtual Keys</strong></p>
<ul>
<li>View deleted keys for audit purposes - <a href="https://github.com/BerriAI/litellm/pull/18228" target="_blank" rel="noopener noreferrer">PR #18228</a>, <a href="https://github.com/BerriAI/litellm/pull/19268" target="_blank" rel="noopener noreferrer">PR #19268</a></li>
<li>Add status query parameter for keys list - <a href="https://github.com/BerriAI/litellm/pull/19260" target="_blank" rel="noopener noreferrer">PR #19260</a></li>
<li>Refetch keys after key creation - <a href="https://github.com/BerriAI/litellm/pull/18994" target="_blank" rel="noopener noreferrer">PR #18994</a></li>
<li>Refresh keys list on delete - <a href="https://github.com/BerriAI/litellm/pull/19262" target="_blank" rel="noopener noreferrer">PR #19262</a></li>
<li>Simplify key generate permission error - <a href="https://github.com/BerriAI/litellm/pull/18997" target="_blank" rel="noopener noreferrer">PR #18997</a></li>
<li>Add search to key edit team dropdown - <a href="https://github.com/BerriAI/litellm/pull/19119" target="_blank" rel="noopener noreferrer">PR #19119</a></li>
</ul>
<p><strong>Teams &amp; Organizations</strong></p>
<ul>
<li>View deleted teams for audit purposes - <a href="https://github.com/BerriAI/litellm/pull/18228" target="_blank" rel="noopener noreferrer">PR #18228</a>, <a href="https://github.com/BerriAI/litellm/pull/19268" target="_blank" rel="noopener noreferrer">PR #19268</a></li>
<li>Add filters to organization table - <a href="https://github.com/BerriAI/litellm/pull/18916" target="_blank" rel="noopener noreferrer">PR #18916</a></li>
<li>Add query parameters to <code>/organization/list</code> - <a href="https://github.com/BerriAI/litellm/pull/18910" target="_blank" rel="noopener noreferrer">PR #18910</a></li>
<li>Add status query parameter for teams list - <a href="https://github.com/BerriAI/litellm/pull/19260" target="_blank" rel="noopener noreferrer">PR #19260</a></li>
<li>Show internal users their spend only - <a href="https://github.com/BerriAI/litellm/pull/19227" target="_blank" rel="noopener noreferrer">PR #19227</a></li>
<li>Allow preventing team admins from deleting members from teams - <a href="https://github.com/BerriAI/litellm/pull/19128" target="_blank" rel="noopener noreferrer">PR #19128</a></li>
<li>Refactor team member icon buttons - <a href="https://github.com/BerriAI/litellm/pull/19192" target="_blank" rel="noopener noreferrer">PR #19192</a></li>
</ul>
<p><strong>Models + Endpoints</strong></p>
<ul>
<li>Display health information in public model hub - <a href="https://github.com/BerriAI/litellm/pull/19256" target="_blank" rel="noopener noreferrer">PR #19256</a>, <a href="https://github.com/BerriAI/litellm/pull/19258" target="_blank" rel="noopener noreferrer">PR #19258</a></li>
<li>Quality of life improvements for Anthropic models - <a href="https://github.com/BerriAI/litellm/pull/19058" target="_blank" rel="noopener noreferrer">PR #19058</a></li>
<li>Create reusable model select component - <a href="https://github.com/BerriAI/litellm/pull/19164" target="_blank" rel="noopener noreferrer">PR #19164</a></li>
<li>Edit settings model dropdown - <a href="https://github.com/BerriAI/litellm/pull/19186" target="_blank" rel="noopener noreferrer">PR #19186</a></li>
<li>Fix model hub client side exception - <a href="https://github.com/BerriAI/litellm/pull/19045" target="_blank" rel="noopener noreferrer">PR #19045</a></li>
</ul>
<p><strong>Usage &amp; Analytics</strong></p>
<ul>
<li>Allow top virtual keys and models to show more entries - <a href="https://github.com/BerriAI/litellm/pull/19050" target="_blank" rel="noopener noreferrer">PR #19050</a></li>
<li>Fix Y axis on model activity chart - <a href="https://github.com/BerriAI/litellm/pull/19055" target="_blank" rel="noopener noreferrer">PR #19055</a></li>
<li>Add Team ID and Team Name in export report - <a href="https://github.com/BerriAI/litellm/pull/19047" target="_blank" rel="noopener noreferrer">PR #19047</a></li>
<li>Add user metrics for Prometheus - <a href="https://github.com/BerriAI/litellm/pull/18785" target="_blank" rel="noopener noreferrer">PR #18785</a></li>
</ul>
<p><strong>SSO &amp; Auth</strong></p>
<ul>
<li>Allow setting custom MSFT Base URLs - <a href="https://github.com/BerriAI/litellm/pull/18977" target="_blank" rel="noopener noreferrer">PR #18977</a></li>
<li>Allow overriding env var attribute names - <a href="https://github.com/BerriAI/litellm/pull/18998" target="_blank" rel="noopener noreferrer">PR #18998</a></li>
<li>Fix SCIM GET /Users error and enforce SCIM 2.0 compliance - <a href="https://github.com/BerriAI/litellm/pull/17420" target="_blank" rel="noopener noreferrer">PR #17420</a></li>
<li>Feature flag for SCIM compliance fix - <a href="https://github.com/BerriAI/litellm/pull/18878" target="_blank" rel="noopener noreferrer">PR #18878</a></li>
</ul>
<p><strong>General UI</strong></p>
<ul>
<li>Add allowClear to dropdown components for better UX - <a href="https://github.com/BerriAI/litellm/pull/18778" target="_blank" rel="noopener noreferrer">PR #18778</a></li>
<li>Add community engagement buttons - <a href="https://github.com/BerriAI/litellm/pull/19114" target="_blank" rel="noopener noreferrer">PR #19114</a></li>
<li>UI Feedback Form - why LiteLLM - <a href="https://github.com/BerriAI/litellm/pull/18999" target="_blank" rel="noopener noreferrer">PR #18999</a></li>
<li>Refactor user and team table filters to reusable component - <a href="https://github.com/BerriAI/litellm/pull/19010" target="_blank" rel="noopener noreferrer">PR #19010</a></li>
<li>Adjusting new badges - <a href="https://github.com/BerriAI/litellm/pull/19278" target="_blank" rel="noopener noreferrer">PR #19278</a></li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs-1">Bugs<a href="https://docs.litellm.ai/release_notes/v1-81-0#bugs-1" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li>Container API routes return 401 for non-admin users - routes missing from openai_routes - <a href="https://github.com/BerriAI/litellm/pull/19115" target="_blank" rel="noopener noreferrer">PR #19115</a></li>
<li>Allow routing to regional endpoints for Containers API - <a href="https://github.com/BerriAI/litellm/pull/19118" target="_blank" rel="noopener noreferrer">PR #19118</a></li>
<li>Fix Azure Storage circular reference error - <a href="https://github.com/BerriAI/litellm/pull/19120" target="_blank" rel="noopener noreferrer">PR #19120</a></li>
<li>Fix prompt deletion fails with Prisma FieldNotFoundError - <a href="https://github.com/BerriAI/litellm/pull/18966" target="_blank" rel="noopener noreferrer">PR #18966</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="ai-integrations">AI Integrations<a href="https://docs.litellm.ai/release_notes/v1-81-0#ai-integrations" class="hash-link" aria-label="Direct link to AI Integrations" title="Direct link to AI Integrations">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="logging">Logging<a href="https://docs.litellm.ai/release_notes/v1-81-0#logging" class="hash-link" aria-label="Direct link to Logging" title="Direct link to Logging">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#opentelemetry">OpenTelemetry</a></strong></p>
<ul>
<li>Update semantic conventions to 1.38 (gen_ai attributes) - <a href="https://github.com/BerriAI/litellm/pull/18793" target="_blank" rel="noopener noreferrer">PR #18793</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#langsmith">LangSmith</a></strong></p>
<ul>
<li>Hoist thread grouping metadata (session_id, thread) - <a href="https://github.com/BerriAI/litellm/pull/18982" target="_blank" rel="noopener noreferrer">PR #18982</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#langfuse">Langfuse</a></strong></p>
<ul>
<li>Include Langfuse logger in JSON logging when Langfuse callback is used - <a href="https://github.com/BerriAI/litellm/pull/19162" target="_blank" rel="noopener noreferrer">PR #19162</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/observability/logfire">Logfire</a></strong></p>
<ul>
<li>Add ability to customize Logfire base URL through env var - <a href="https://github.com/BerriAI/litellm/pull/19148" target="_blank" rel="noopener noreferrer">PR #19148</a></li>
</ul>
</li>
<li>
<p><strong>General Logging</strong></p>
<ul>
<li>Enable JSON logging via configuration and add regression test - <a href="https://github.com/BerriAI/litellm/pull/19037" target="_blank" rel="noopener noreferrer">PR #19037</a></li>
<li>Fix header forwarding for embeddings endpoint - <a href="https://github.com/BerriAI/litellm/pull/18960" target="_blank" rel="noopener noreferrer">PR #18960</a></li>
<li>Preserve llm_provider-* headers in error responses - <a href="https://github.com/BerriAI/litellm/pull/19020" target="_blank" rel="noopener noreferrer">PR #19020</a></li>
<li>Fix turn_off_message_logging not redacting request messages in proxy_server_request field - <a href="https://github.com/BerriAI/litellm/pull/18897" target="_blank" rel="noopener noreferrer">PR #18897</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails">Guardrails<a href="https://docs.litellm.ai/release_notes/v1-81-0#guardrails" class="hash-link" aria-label="Direct link to Guardrails" title="Direct link to Guardrails">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/grayswan">Grayswan</a></strong></p>
<ul>
<li>Implement fail-open option (default: True) - <a href="https://github.com/BerriAI/litellm/pull/18266" target="_blank" rel="noopener noreferrer">PR #18266</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/pangea">Pangea</a></strong></p>
<ul>
<li>Respect <code>default_on</code> during initialization - <a href="https://github.com/BerriAI/litellm/pull/18912" target="_blank" rel="noopener noreferrer">PR #18912</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/panw_prisma_airs">Panw Prisma AIRS</a></strong></p>
<ul>
<li>Add custom violation message support - <a href="https://github.com/BerriAI/litellm/pull/19272" target="_blank" rel="noopener noreferrer">PR #19272</a></li>
</ul>
</li>
<li>
<p><strong>General Guardrails</strong></p>
<ul>
<li>Fix SerializationIterator error and pass tools to guardrail - <a href="https://github.com/BerriAI/litellm/pull/18932" target="_blank" rel="noopener noreferrer">PR #18932</a></li>
<li>Properly handle custom guardrails parameters - <a href="https://github.com/BerriAI/litellm/pull/18978" target="_blank" rel="noopener noreferrer">PR #18978</a></li>
<li>Use clean error messages for blocked requests - <a href="https://github.com/BerriAI/litellm/pull/19023" target="_blank" rel="noopener noreferrer">PR #19023</a></li>
<li>Guardrail moderation support with responses API - <a href="https://github.com/BerriAI/litellm/pull/18957" target="_blank" rel="noopener noreferrer">PR #18957</a></li>
<li>Fix model-level guardrails not taking effect - <a href="https://github.com/BerriAI/litellm/pull/18895" target="_blank" rel="noopener noreferrer">PR #18895</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-81-0#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li>
<p><strong>Cost Calculation Fixes</strong></p>
<ul>
<li>Include IMAGE token count in cost calculation for Gemini models - <a href="https://github.com/BerriAI/litellm/pull/18876" target="_blank" rel="noopener noreferrer">PR #18876</a></li>
<li>Fix negative text_tokens when using cache with images - <a href="https://github.com/BerriAI/litellm/pull/18768" target="_blank" rel="noopener noreferrer">PR #18768</a></li>
<li>Fix image tokens spend logging for <code>/images/generations</code> - <a href="https://github.com/BerriAI/litellm/pull/19009" target="_blank" rel="noopener noreferrer">PR #19009</a></li>
<li>Fix incorrect <code>prompt_tokens_details</code> in Gemini Image Generation - <a href="https://github.com/BerriAI/litellm/pull/19070" target="_blank" rel="noopener noreferrer">PR #19070</a></li>
<li>Fix case-insensitive model cost map lookup - <a href="https://github.com/BerriAI/litellm/pull/18208" target="_blank" rel="noopener noreferrer">PR #18208</a></li>
</ul>
</li>
<li>
<p><strong>Pricing Updates</strong></p>
<ul>
<li>Correct pricing for <code>openrouter/openai/gpt-oss-20b</code> - <a href="https://github.com/BerriAI/litellm/pull/18899" target="_blank" rel="noopener noreferrer">PR #18899</a></li>
<li>Add pricing for <code>azure_ai/claude-opus-4-5</code> - <a href="https://github.com/BerriAI/litellm/pull/19003" target="_blank" rel="noopener noreferrer">PR #19003</a></li>
<li>Update Novita models prices - <a href="https://github.com/BerriAI/litellm/pull/19005" target="_blank" rel="noopener noreferrer">PR #19005</a></li>
<li>Fix Azure Grok prices - <a href="https://github.com/BerriAI/litellm/pull/19102" target="_blank" rel="noopener noreferrer">PR #19102</a></li>
<li>Fix GCP GLM-4.7 pricing - <a href="https://github.com/BerriAI/litellm/pull/19172" target="_blank" rel="noopener noreferrer">PR #19172</a></li>
<li>Sync DeepSeek chat/reasoner to V3.2 pricing - <a href="https://github.com/BerriAI/litellm/pull/18884" target="_blank" rel="noopener noreferrer">PR #18884</a></li>
<li>Correct cache_read pricing for gemini-2.5-pro models - <a href="https://github.com/BerriAI/litellm/pull/18157" target="_blank" rel="noopener noreferrer">PR #18157</a></li>
</ul>
</li>
<li>
<p><strong>Budget &amp; Rate Limiting</strong></p>
<ul>
<li>Correct budget limit validation operator (&gt;=) for team members - <a href="https://github.com/BerriAI/litellm/pull/19207" target="_blank" rel="noopener noreferrer">PR #19207</a></li>
<li>Fix TPM 25% limiting by ensuring priority queue logic - <a href="https://github.com/BerriAI/litellm/pull/19092" target="_blank" rel="noopener noreferrer">PR #19092</a></li>
<li>Cleanup spend logs cron verification, fix, and docs - <a href="https://github.com/BerriAI/litellm/pull/19085" target="_blank" rel="noopener noreferrer">PR #19085</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-gateway">MCP Gateway<a href="https://docs.litellm.ai/release_notes/v1-81-0#mcp-gateway" class="hash-link" aria-label="Direct link to MCP Gateway" title="Direct link to MCP Gateway">​</a></h2>
<ul>
<li>Prevent duplicate MCP reload scheduler registration - <a href="https://github.com/BerriAI/litellm/pull/18934" target="_blank" rel="noopener noreferrer">PR #18934</a></li>
<li>Forward MCP extra headers case-insensitively - <a href="https://github.com/BerriAI/litellm/pull/18940" target="_blank" rel="noopener noreferrer">PR #18940</a></li>
<li>Fix MCP REST auth checks - <a href="https://github.com/BerriAI/litellm/pull/19051" target="_blank" rel="noopener noreferrer">PR #19051</a></li>
<li>Fix generating two telemetry events in responses - <a href="https://github.com/BerriAI/litellm/pull/18938" target="_blank" rel="noopener noreferrer">PR #18938</a></li>
<li>Fix MCP chat completions - <a href="https://github.com/BerriAI/litellm/pull/19129" target="_blank" rel="noopener noreferrer">PR #19129</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements">Performance / Loadbalancing / Reliability improvements<a href="https://docs.litellm.ai/release_notes/v1-81-0#performance--loadbalancing--reliability-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements" title="Direct link to Performance / Loadbalancing / Reliability improvements">​</a></h2>
<ul>
<li>
<p><strong>Performance Improvements</strong></p>
<ul>
<li>Remove bottleneck causing high CPU usage &amp; overhead under heavy load - <a href="https://github.com/BerriAI/litellm/pull/19049" target="_blank" rel="noopener noreferrer">PR #19049</a></li>
<li>Add CI enforcement for O(1) operations in <code>_get_model_cost_key</code> to prevent performance regressions - <a href="https://github.com/BerriAI/litellm/pull/19052" target="_blank" rel="noopener noreferrer">PR #19052</a></li>
<li>Fix Azure embeddings JSON parsing to prevent connection leaks and ensure proper router cooldown - <a href="https://github.com/BerriAI/litellm/pull/19167" target="_blank" rel="noopener noreferrer">PR #19167</a></li>
<li>Do not fallback to token counter if <code>disable_token_counter</code> is enabled - <a href="https://github.com/BerriAI/litellm/pull/19041" target="_blank" rel="noopener noreferrer">PR #19041</a></li>
</ul>
</li>
<li>
<p><strong>Reliability</strong></p>
<ul>
<li>Add fallback endpoints support - <a href="https://github.com/BerriAI/litellm/pull/19185" target="_blank" rel="noopener noreferrer">PR #19185</a></li>
<li>Fix stream_timeout parameter functionality - <a href="https://github.com/BerriAI/litellm/pull/19191" target="_blank" rel="noopener noreferrer">PR #19191</a></li>
<li>Fix model matching priority in configuration - <a href="https://github.com/BerriAI/litellm/pull/19012" target="_blank" rel="noopener noreferrer">PR #19012</a></li>
<li>Fix num_retries in litellm_params as per config - <a href="https://github.com/BerriAI/litellm/pull/18975" target="_blank" rel="noopener noreferrer">PR #18975</a></li>
<li>Handle exceptions without response parameter - <a href="https://github.com/BerriAI/litellm/pull/18919" target="_blank" rel="noopener noreferrer">PR #18919</a></li>
</ul>
</li>
<li>
<p><strong>Infrastructure</strong></p>
<ul>
<li>Add Custom CA certificates to boto3 clients - <a href="https://github.com/BerriAI/litellm/pull/18942" target="_blank" rel="noopener noreferrer">PR #18942</a></li>
<li>Update boto3 to 1.40.15 and aioboto3 to 15.5.0 - <a href="https://github.com/BerriAI/litellm/pull/19090" target="_blank" rel="noopener noreferrer">PR #19090</a></li>
<li>Make keepalive_timeout parameter work for Gunicorn - <a href="https://github.com/BerriAI/litellm/pull/19087" target="_blank" rel="noopener noreferrer">PR #19087</a></li>
</ul>
</li>
<li>
<p><strong>Helm Chart</strong></p>
<ul>
<li>Fix mount config.yaml as single file in Helm chart - <a href="https://github.com/BerriAI/litellm/pull/19146" target="_blank" rel="noopener noreferrer">PR #19146</a></li>
<li>Sync Helm chart versioning with production standards and Docker versions - <a href="https://github.com/BerriAI/litellm/pull/18868" target="_blank" rel="noopener noreferrer">PR #18868</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="database-changes">Database Changes<a href="https://docs.litellm.ai/release_notes/v1-81-0#database-changes" class="hash-link" aria-label="Direct link to Database Changes" title="Direct link to Database Changes">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="schema-updates">Schema Updates<a href="https://docs.litellm.ai/release_notes/v1-81-0#schema-updates" class="hash-link" aria-label="Direct link to Schema Updates" title="Direct link to Schema Updates">​</a></h3>
<table><thead><tr><th>Table</th><th>Change Type</th><th>Description</th><th>PR</th></tr></thead><tbody><tr><td><code>LiteLLM_ProxyModelTable</code></td><td>New Columns</td><td>Added <code>created_at</code> and <code>updated_at</code> timestamp fields</td><td><a href="https://github.com/BerriAI/litellm/pull/18937" target="_blank" rel="noopener noreferrer">PR #18937</a></td></tr></tbody></table>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates">Documentation Updates<a href="https://docs.litellm.ai/release_notes/v1-81-0#documentation-updates" class="hash-link" aria-label="Direct link to Documentation Updates" title="Direct link to Documentation Updates">​</a></h2>
<ul>
<li>Add LiteLLM architecture md doc - <a href="https://github.com/BerriAI/litellm/pull/19057" target="_blank" rel="noopener noreferrer">PR #19057</a>, <a href="https://github.com/BerriAI/litellm/pull/19252" target="_blank" rel="noopener noreferrer">PR #19252</a></li>
<li>Add troubleshooting guide - <a href="https://github.com/BerriAI/litellm/pull/19096" target="_blank" rel="noopener noreferrer">PR #19096</a>, <a href="https://github.com/BerriAI/litellm/pull/19097" target="_blank" rel="noopener noreferrer">PR #19097</a>, <a href="https://github.com/BerriAI/litellm/pull/19099" target="_blank" rel="noopener noreferrer">PR #19099</a></li>
<li>Add structured issue reporting guides for CPU and memory issues - <a href="https://github.com/BerriAI/litellm/pull/19117" target="_blank" rel="noopener noreferrer">PR #19117</a></li>
<li>Add Redis requirement warning for high-traffic deployments - <a href="https://github.com/BerriAI/litellm/pull/18892" target="_blank" rel="noopener noreferrer">PR #18892</a></li>
<li>Update load balancing and routing with enable_pre_call_checks - <a href="https://github.com/BerriAI/litellm/pull/18888" target="_blank" rel="noopener noreferrer">PR #18888</a></li>
<li>Updated pass_through with guided param - <a href="https://github.com/BerriAI/litellm/pull/18886" target="_blank" rel="noopener noreferrer">PR #18886</a></li>
<li>Update message content types link and add content types table - <a href="https://github.com/BerriAI/litellm/pull/18209" target="_blank" rel="noopener noreferrer">PR #18209</a></li>
<li>Add Redis initialization with kwargs - <a href="https://github.com/BerriAI/litellm/pull/19183" target="_blank" rel="noopener noreferrer">PR #19183</a></li>
<li>Improve documentation for routing LLM calls via SAP Gen AI Hub - <a href="https://github.com/BerriAI/litellm/pull/19166" target="_blank" rel="noopener noreferrer">PR #19166</a></li>
<li>Deleted Keys and Teams docs - <a href="https://github.com/BerriAI/litellm/pull/19291" target="_blank" rel="noopener noreferrer">PR #19291</a></li>
<li>Claude Code end user tracking guide - <a href="https://github.com/BerriAI/litellm/pull/19176" target="_blank" rel="noopener noreferrer">PR #19176</a></li>
<li>Add MCP troubleshooting guide - <a href="https://github.com/BerriAI/litellm/pull/19122" target="_blank" rel="noopener noreferrer">PR #19122</a></li>
<li>Add auth message UI documentation - <a href="https://github.com/BerriAI/litellm/pull/19063" target="_blank" rel="noopener noreferrer">PR #19063</a></li>
<li>Add guide for mounting custom callbacks in Helm/K8s - <a href="https://github.com/BerriAI/litellm/pull/19136" target="_blank" rel="noopener noreferrer">PR #19136</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes-1">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-81-0#bug-fixes-1" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h2>
<ul>
<li>Fix Swagger UI path execute error with server_root_path in OpenAPI schema - <a href="https://github.com/BerriAI/litellm/pull/18947" target="_blank" rel="noopener noreferrer">PR #18947</a></li>
<li>Normalize OpenAI SDK BaseModel choices/messages to avoid Pydantic serializer warnings - <a href="https://github.com/BerriAI/litellm/pull/18972" target="_blank" rel="noopener noreferrer">PR #18972</a></li>
<li>Add contextual gap checks and word-form digits - <a href="https://github.com/BerriAI/litellm/pull/18301" target="_blank" rel="noopener noreferrer">PR #18301</a></li>
<li>Clean up orphaned files from repository root - <a href="https://github.com/BerriAI/litellm/pull/19150" target="_blank" rel="noopener noreferrer">PR #19150</a></li>
<li>Include proxy/prisma_migration.py in non-root - <a href="https://github.com/BerriAI/litellm/pull/18971" target="_blank" rel="noopener noreferrer">PR #18971</a></li>
<li>Update prisma_migration.py - <a href="https://github.com/BerriAI/litellm/pull/19083" target="_blank" rel="noopener noreferrer">PR #19083</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-81-0#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@yogeshwaran10 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18898" target="_blank" rel="noopener noreferrer">PR #18898</a></li>
<li>@theonlypal made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18937" target="_blank" rel="noopener noreferrer">PR #18937</a></li>
<li>@jonmagic made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18935" target="_blank" rel="noopener noreferrer">PR #18935</a></li>
<li>@houdataali made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19025" target="_blank" rel="noopener noreferrer">PR #19025</a></li>
<li>@hummat made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18972" target="_blank" rel="noopener noreferrer">PR #18972</a></li>
<li>@berkeyalciin made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18966" target="_blank" rel="noopener noreferrer">PR #18966</a></li>
<li>@MateuszOssGit made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18959" target="_blank" rel="noopener noreferrer">PR #18959</a></li>
<li>@xfan001 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18947" target="_blank" rel="noopener noreferrer">PR #18947</a></li>
<li>@nulone made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18884" target="_blank" rel="noopener noreferrer">PR #18884</a></li>
<li>@debnil-mercor made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18919" target="_blank" rel="noopener noreferrer">PR #18919</a></li>
<li>@hakhundov made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17420" target="_blank" rel="noopener noreferrer">PR #17420</a></li>
<li>@rohanwinsor made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19078" target="_blank" rel="noopener noreferrer">PR #19078</a></li>
<li>@pgolm made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19020" target="_blank" rel="noopener noreferrer">PR #19020</a></li>
<li>@vikigenius made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19148" target="_blank" rel="noopener noreferrer">PR #19148</a></li>
<li>@burnerburnerburnerman made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19090" target="_blank" rel="noopener noreferrer">PR #19090</a></li>
<li>@yfge made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19076" target="_blank" rel="noopener noreferrer">PR #19076</a></li>
<li>@danielnyari-seon made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19083" target="_blank" rel="noopener noreferrer">PR #19083</a></li>
<li>@guilherme-segantini made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19166" target="_blank" rel="noopener noreferrer">PR #19166</a></li>
<li>@jgreek made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19147" target="_blank" rel="noopener noreferrer">PR #19147</a></li>
<li>@anand-kamble made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19193" target="_blank" rel="noopener noreferrer">PR #19193</a></li>
<li>@neubig made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/19162" target="_blank" rel="noopener noreferrer">PR #19162</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog">Full Changelog<a href="https://docs.litellm.ai/release_notes/v1-81-0#full-changelog" class="hash-link" aria-label="Direct link to Full Changelog" title="Direct link to Full Changelog">​</a></h2>
<p><strong><a href="https://github.com/BerriAI/litellm/compare/v1.80.15.rc.1...v1.81.0.rc.1" target="_blank" rel="noopener noreferrer">View complete changelog on GitHub</a></strong></p>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[v1.80.15-stable - Manus API Support]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-80-15</id>
        <link href="https://docs.litellm.ai/release_notes/v1-80-15"/>
        <updated>2026-01-10T10:00:00.000Z</updated>
        <summary type="html"><![CDATA[Deploy this version]]></summary>
        <content type="html"><![CDATA[<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-80-15#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">docker.litellm.ai/berriai/litellm:v1.80.15-stable.1</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.80.15</span><br></span></code></pre></div></div></div></div></div>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-80-15#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<ul>
<li><strong>Manus API Support</strong> - <a href="https://docs.litellm.ai/docs/providers/manus">New provider support for Manus API on /responses and GET /responses endpoints</a></li>
<li><strong>MiniMax Provider</strong> - <a href="https://docs.litellm.ai/docs/providers/minimax">Full support for MiniMax chat completions, TTS, and Anthropic native endpoint</a></li>
<li><strong>AWS Polly TTS</strong> - <a href="https://docs.litellm.ai/docs/providers/aws_polly">New TTS provider using AWS Polly API</a></li>
<li><strong>SSO Role Mapping</strong> - Configure role mappings for SSO providers directly in the UI</li>
<li><strong>Cost Estimator</strong> - New UI tool for estimating costs across multiple models and requests</li>
<li><strong>MCP Global Mode</strong> - <a href="https://docs.litellm.ai/docs/mcp">Configure MCP servers globally with visibility controls</a></li>
<li><strong>Interactions API Bridge</strong> - <a href="https://docs.litellm.ai/docs/interactions">Use all LiteLLM providers with the Interactions API</a></li>
<li><strong>RAG Query Endpoint</strong> - <a href="https://docs.litellm.ai/docs/search/index">New RAG Search/Query endpoint for retrieval-augmented generation</a></li>
<li><strong>UI Usage - Endpoint Activity</strong> - <a href="https://docs.litellm.ai/docs/proxy/endpoint_activity.md">Users can now see Endpoint Activity Metrics in the UI</a></li>
<li><strong>50% Overhead Reduction</strong> - LiteLLM now sends 2.5× more requests to LLM providers</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance---50-overhead-reduction">Performance - 50% Overhead Reduction<a href="https://docs.litellm.ai/release_notes/v1-80-15#performance---50-overhead-reduction" class="hash-link" aria-label="Direct link to Performance - 50% Overhead Reduction" title="Direct link to Performance - 50% Overhead Reduction">​</a></h2>
<p>LiteLLM now sends 2.5× more requests to LLM providers by replacing sequential if/elif chains with O(1) dictionary lookups for provider configuration resolution (92.7% faster). This optimization has a high impact because it runs inside the client decorator, which is invoked on every HTTP request made to the proxy server.</p>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="before">Before<a href="https://docs.litellm.ai/release_notes/v1-80-15#before" class="hash-link" aria-label="Direct link to Before" title="Direct link to Before">​</a></h3>
<blockquote>
<p><strong>Note:</strong> Worse-looking provider metrics are a good sign here—they indicate requests spend less time inside LiteLLM.</p>
</blockquote>
<div class="language-text codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-text codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">============================================================</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">Fake LLM Provider Stats (When called by LiteLLM)</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">============================================================</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">Total Time:            0.56s</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">Requests/Second:       10746.68</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain" style="display:inline-block"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">Latency Statistics (seconds):</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">   Mean:               0.2039s</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">   Median (p50):       0.2310s</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">   Min:                0.0323s</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">   Max:                0.3928s</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">   Std Dev:            0.1166s</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">   p95:                0.3574s</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">   p99:                0.3748s</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain" style="display:inline-block"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">Status Codes:</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">   200: 6000</span><br></span></code></pre></div></div>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="after">After<a href="https://docs.litellm.ai/release_notes/v1-80-15#after" class="hash-link" aria-label="Direct link to After" title="Direct link to After">​</a></h3>
<div class="language-text codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-text codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">============================================================</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">Fake LLM Provider Stats (When called by LiteLLM)</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">============================================================</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">Total Time:            1.42s</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">Requests/Second:       4224.49</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain" style="display:inline-block"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">Latency Statistics (seconds):</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">   Mean:               0.5300s</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">   Median (p50):       0.5871s</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">   Min:                0.0885s</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">   Max:                1.0482s</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">   Std Dev:            0.3065s</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">   p95:                0.9750s</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">   p99:                1.0444s</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain" style="display:inline-block"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">Status Codes:</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">   200: 6000</span><br></span></code></pre></div></div>
<blockquote>
<p>The benchmarks run LiteLLM locally with a lightweight LLM provider to eliminate network latency, isolating internal overhead and bottlenecks so we can focus on reducing pure LiteLLM overhead on a single instance.</p>
</blockquote>
<hr>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="ui-usage---endpoint-activity">UI Usage - Endpoint Activity<a href="https://docs.litellm.ai/release_notes/v1-80-15#ui-usage---endpoint-activity" class="hash-link" aria-label="Direct link to UI Usage - Endpoint Activity" title="Direct link to UI Usage - Endpoint Activity">​</a></h3>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAFCAYAAAB8ZH1oAAAACXBIWXMAACxLAAAsSwGlPZapAAAAnElEQVR4nCWLOQ7CQBAE/XVSXkEEGR+AhJzMRoIEbAQJl8HsNTvrLWTTUkstVXXRNGfqusZah/eeKBFVRaMiEgnBY0ygKMuSqqr+ogRa9x3FoEJUJSWlbYXCGIO1Fu88mhMv6UaYc6bPPSmlscXwTqoYZ3l8n0gfGbK8bNh9TuPWQQwSIMG+a5ge5izOa+7+zWQ/Y3XbcnUPqs+RH8/uvaSS/Yj7AAAAAElFTkSuQmCC&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="334"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/ui_endpoint_activity.412e94a.640.png" srcset="/assets/ideal-img/ui_endpoint_activity.412e94a.640.png 640w,/assets/ideal-img/ui_endpoint_activity.1d30413.1920.png 1920w" width="640" height="334"></noscript></div>
<p>Users can now see Endpoint Activity Metrics in the UI.</p>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-providers-and-endpoints">New Providers and Endpoints<a href="https://docs.litellm.ai/release_notes/v1-80-15#new-providers-and-endpoints" class="hash-link" aria-label="Direct link to New Providers and Endpoints" title="Direct link to New Providers and Endpoints">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="new-providers-11-new-providers">New Providers (11 new providers)<a href="https://docs.litellm.ai/release_notes/v1-80-15#new-providers-11-new-providers" class="hash-link" aria-label="Direct link to New Providers (11 new providers)" title="Direct link to New Providers (11 new providers)">​</a></h3>
<table><thead><tr><th>Provider</th><th>Supported LiteLLM Endpoints</th><th>Description</th></tr></thead><tbody><tr><td><a href="https://docs.litellm.ai/docs/providers/manus">Manus</a></td><td><code>/responses</code></td><td>Manus API for agentic workflows</td></tr><tr><td><a href="https://docs.litellm.ai/docs/providers/manus">Manus</a></td><td><code>GET /responses</code></td><td>Manus API for retrieving responses</td></tr><tr><td><a href="https://docs.litellm.ai/docs/providers/manus">Manus</a></td><td><code>/files</code></td><td>Manus API for file management</td></tr><tr><td><a href="https://docs.litellm.ai/docs/providers/minimax">MiniMax</a></td><td><code>/chat/completions</code></td><td>MiniMax chat completions</td></tr><tr><td><a href="https://docs.litellm.ai/docs/providers/minimax">MiniMax</a></td><td><code>/audio/speech</code></td><td>MiniMax text-to-speech</td></tr><tr><td><a href="https://docs.litellm.ai/docs/providers/aws_polly">AWS Polly</a></td><td><code>/audio/speech</code></td><td>AWS Polly text-to-speech API</td></tr><tr><td><a href="https://docs.litellm.ai/docs/providers/gigachat">GigaChat</a></td><td><code>/chat/completions</code></td><td>GigaChat provider for Russian language AI</td></tr><tr><td><a href="https://docs.litellm.ai/docs/providers/llamagate">LlamaGate</a></td><td><code>/chat/completions</code></td><td>LlamaGate chat completions</td></tr><tr><td><a href="https://docs.litellm.ai/docs/providers/llamagate">LlamaGate</a></td><td><code>/embeddings</code></td><td>LlamaGate embeddings</td></tr><tr><td><a href="https://docs.litellm.ai/docs/providers/abliteration">Abliteration AI</a></td><td><code>/chat/completions</code></td><td>Abliteration.ai provider support</td></tr><tr><td><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></td><td><code>/v1/messages/count_tokens</code></td><td>Bedrock as new provider for token counting</td></tr></tbody></table>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="new-llm-api-endpoints-3-new-endpoints">New LLM API Endpoints (3 new endpoints)<a href="https://docs.litellm.ai/release_notes/v1-80-15#new-llm-api-endpoints-3-new-endpoints" class="hash-link" aria-label="Direct link to New LLM API Endpoints (3 new endpoints)" title="Direct link to New LLM API Endpoints (3 new endpoints)">​</a></h3>
<table><thead><tr><th>Endpoint</th><th>Method</th><th>Description</th><th>Documentation</th></tr></thead><tbody><tr><td><code>/responses/compact</code></td><td>POST</td><td>Compact responses API endpoint</td><td><a href="https://docs.litellm.ai/docs/response_api">Docs</a></td></tr><tr><td><code>/rag/query</code></td><td>POST</td><td>RAG Search/Query endpoint</td><td><a href="https://docs.litellm.ai/docs/search/index">Docs</a></td></tr><tr><td><code>/containers/{id}/files</code></td><td>POST</td><td>Upload files to containers</td><td><a href="https://docs.litellm.ai/docs/container_files">Docs</a></td></tr></tbody></table>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-80-15#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support-100-new-models">New Model Support (100+ new models)<a href="https://docs.litellm.ai/release_notes/v1-80-15#new-model-support-100-new-models" class="hash-link" aria-label="Direct link to New Model Support (100+ new models)" title="Direct link to New Model Support (100+ new models)">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th><th>Features</th></tr></thead><tbody><tr><td>Azure</td><td><code>azure/gpt-5.2</code></td><td>400K</td><td>$1.75</td><td>$14.00</td><td>Reasoning, vision, caching</td></tr><tr><td>Azure</td><td><code>azure/gpt-5.2-chat</code></td><td>128K</td><td>$1.75</td><td>$14.00</td><td>Reasoning, vision</td></tr><tr><td>Azure</td><td><code>azure/gpt-5.2-pro</code></td><td>400K</td><td>$21.00</td><td>$168.00</td><td>Reasoning, vision, web search</td></tr><tr><td>Azure</td><td><code>azure/gpt-image-1.5</code></td><td>-</td><td>Token-based</td><td>Token-based</td><td>Image generation/editing</td></tr><tr><td>Azure AI</td><td><code>azure_ai/gpt-oss-120b</code></td><td>131K</td><td>$0.15</td><td>$0.60</td><td>Function calling</td></tr><tr><td>Azure AI</td><td><code>azure_ai/flux.2-pro</code></td><td>-</td><td>-</td><td>$0.04/image</td><td>Image generation</td></tr><tr><td>Azure AI</td><td><code>azure_ai/deepseek-v3.2</code></td><td>164K</td><td>$0.58</td><td>$1.68</td><td>Reasoning, function calling</td></tr><tr><td>Bedrock</td><td><code>amazon.nova-2-multimodal-embeddings-v1:0</code></td><td>8K</td><td>$0.135</td><td>-</td><td>Multimodal embeddings</td></tr><tr><td>Bedrock</td><td><code>writer.palmyra-x4-v1:0</code></td><td>128K</td><td>$2.50</td><td>$10.00</td><td>Function calling, PDF</td></tr><tr><td>Bedrock</td><td><code>writer.palmyra-x5-v1:0</code></td><td>1M</td><td>$0.60</td><td>$6.00</td><td>Function calling, PDF</td></tr><tr><td>Bedrock</td><td><code>moonshot.kimi-k2-v1:0</code></td><td>-</td><td>-</td><td>-</td><td>Kimi K2 model</td></tr><tr><td>Cerebras</td><td><code>cerebras/zai-glm-4.6</code></td><td>128K</td><td>$2.25</td><td>$2.75</td><td>Reasoning, function calling</td></tr><tr><td>GigaChat</td><td><code>gigachat/GigaChat-2-Lite</code></td><td>-</td><td>-</td><td>-</td><td>Chat completions</td></tr><tr><td>GigaChat</td><td><code>gigachat/GigaChat-2-Max</code></td><td>-</td><td>-</td><td>-</td><td>Chat completions</td></tr><tr><td>GigaChat</td><td><code>gigachat/GigaChat-2-Pro</code></td><td>-</td><td>-</td><td>-</td><td>Chat completions</td></tr><tr><td>Gemini</td><td><code>gemini/veo-3.1-generate-001</code></td><td>-</td><td>-</td><td>-</td><td>Video generation</td></tr><tr><td>Gemini</td><td><code>gemini/veo-3.1-fast-generate-001</code></td><td>-</td><td>-</td><td>-</td><td>Video generation</td></tr><tr><td>GitHub Copilot</td><td>25+ models</td><td>Various</td><td>-</td><td>-</td><td>Chat completions</td></tr><tr><td>LlamaGate</td><td>15+ models</td><td>Various</td><td>-</td><td>-</td><td>Chat, vision, embeddings</td></tr><tr><td>MiniMax</td><td><code>minimax/abab7-chat-preview</code></td><td>-</td><td>-</td><td>-</td><td>Chat completions</td></tr><tr><td>Novita</td><td>80+ models</td><td>Various</td><td>Various</td><td>Various</td><td>Chat, vision, embeddings</td></tr><tr><td>OpenRouter</td><td><code>openrouter/google/gemini-3-flash-preview</code></td><td>-</td><td>-</td><td>-</td><td>Chat completions</td></tr><tr><td>Together AI</td><td>Multiple models</td><td>Various</td><td>Various</td><td>Various</td><td>Response schema support</td></tr><tr><td>Vertex AI</td><td><code>vertex_ai/zai-glm-4.7</code></td><td>-</td><td>-</td><td>-</td><td>GLM 4.7 support</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-80-15#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Gemini</a></strong>
<ul>
<li>Add image tokens in chat completion - <a href="https://github.com/BerriAI/litellm/pull/18327" target="_blank" rel="noopener noreferrer">PR #18327</a></li>
<li>Add usage object in image generation - <a href="https://github.com/BerriAI/litellm/pull/18328" target="_blank" rel="noopener noreferrer">PR #18328</a></li>
<li>Add thought signature support via tool call id - <a href="https://github.com/BerriAI/litellm/pull/18374" target="_blank" rel="noopener noreferrer">PR #18374</a></li>
<li>Add thought signature for non tool call requests - <a href="https://github.com/BerriAI/litellm/pull/18581" target="_blank" rel="noopener noreferrer">PR #18581</a></li>
<li>Preserve system instructions - <a href="https://github.com/BerriAI/litellm/pull/18585" target="_blank" rel="noopener noreferrer">PR #18585</a></li>
<li>Fix Gemini 3 images in tool response - <a href="https://github.com/BerriAI/litellm/pull/18190" target="_blank" rel="noopener noreferrer">PR #18190</a></li>
<li>Support snake_case for google_search tool parameters - <a href="https://github.com/BerriAI/litellm/pull/18451" target="_blank" rel="noopener noreferrer">PR #18451</a></li>
<li>Google GenAI adapter inline data support - <a href="https://github.com/BerriAI/litellm/pull/18477" target="_blank" rel="noopener noreferrer">PR #18477</a></li>
<li>Add deprecation_date for discontinued Google models - <a href="https://github.com/BerriAI/litellm/pull/18550" target="_blank" rel="noopener noreferrer">PR #18550</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/vertex">Vertex AI</a></strong>
<ul>
<li>Add centralized get_vertex_base_url() helper for global location support - <a href="https://github.com/BerriAI/litellm/pull/18410" target="_blank" rel="noopener noreferrer">PR #18410</a></li>
<li>Convert image URLs to base64 for Vertex AI Anthropic - <a href="https://github.com/BerriAI/litellm/pull/18497" target="_blank" rel="noopener noreferrer">PR #18497</a></li>
<li>Separate Tool objects for each tool type per API spec - <a href="https://github.com/BerriAI/litellm/pull/18514" target="_blank" rel="noopener noreferrer">PR #18514</a></li>
<li>Add thought_signatures to VertexGeminiConfig - <a href="https://github.com/BerriAI/litellm/pull/18853" target="_blank" rel="noopener noreferrer">PR #18853</a></li>
<li>Add support for Vertex AI API keys - <a href="https://github.com/BerriAI/litellm/pull/18806" target="_blank" rel="noopener noreferrer">PR #18806</a></li>
<li>Add zai glm-4.7 model support - <a href="https://github.com/BerriAI/litellm/pull/18782" target="_blank" rel="noopener noreferrer">PR #18782</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/azure/azure">Azure</a></strong>
<ul>
<li>Add Azure gpt-image-1.5 pricing to cost map - <a href="https://github.com/BerriAI/litellm/pull/18347" target="_blank" rel="noopener noreferrer">PR #18347</a></li>
<li>Add azure/gpt-5.2-chat model - <a href="https://github.com/BerriAI/litellm/pull/18361" target="_blank" rel="noopener noreferrer">PR #18361</a></li>
<li>Add support for image generation via Azure AD token - <a href="https://github.com/BerriAI/litellm/pull/18413" target="_blank" rel="noopener noreferrer">PR #18413</a></li>
<li>Add logprobs support for Azure OpenAI GPT-5.2 model - <a href="https://github.com/BerriAI/litellm/pull/18856" target="_blank" rel="noopener noreferrer">PR #18856</a></li>
<li>Add Azure BFL Flux 2 models for image generation and editing - <a href="https://github.com/BerriAI/litellm/pull/18764" target="_blank" rel="noopener noreferrer">PR #18764</a>, <a href="https://github.com/BerriAI/litellm/pull/18766" target="_blank" rel="noopener noreferrer">PR #18766</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong>
<ul>
<li>Add Bedrock Kimi K2 model support - <a href="https://github.com/BerriAI/litellm/pull/18797" target="_blank" rel="noopener noreferrer">PR #18797</a></li>
<li>Add support for model id in bedrock passthrough - <a href="https://github.com/BerriAI/litellm/pull/18800" target="_blank" rel="noopener noreferrer">PR #18800</a></li>
<li>Fix Nova model detection for Bedrock provider - <a href="https://github.com/BerriAI/litellm/pull/18250" target="_blank" rel="noopener noreferrer">PR #18250</a></li>
<li>Ensure toolUse.input is always a dict when converting from OpenAI format - <a href="https://github.com/BerriAI/litellm/pull/18414" target="_blank" rel="noopener noreferrer">PR #18414</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/databricks">Databricks</a></strong>
<ul>
<li>Add enhanced authentication, security features, and custom user-agent support - <a href="https://github.com/BerriAI/litellm/pull/18349" target="_blank" rel="noopener noreferrer">PR #18349</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/minimax">MiniMax</a></strong>
<ul>
<li>Add MiniMax chat completion support - <a href="https://github.com/BerriAI/litellm/pull/18380" target="_blank" rel="noopener noreferrer">PR #18380</a></li>
<li>Add Anthropic native endpoint support for MiniMax - <a href="https://github.com/BerriAI/litellm/pull/18377" target="_blank" rel="noopener noreferrer">PR #18377</a></li>
<li>Add support for MiniMax TTS - <a href="https://github.com/BerriAI/litellm/pull/18334" target="_blank" rel="noopener noreferrer">PR #18334</a></li>
<li>Add MiniMax provider support to UI dashboard - <a href="https://github.com/BerriAI/litellm/pull/18496" target="_blank" rel="noopener noreferrer">PR #18496</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/togetherai">Together AI</a></strong>
<ul>
<li>Add supports_response_schema to all supported Together AI models - <a href="https://github.com/BerriAI/litellm/pull/18368" target="_blank" rel="noopener noreferrer">PR #18368</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/openrouter">OpenRouter</a></strong>
<ul>
<li>Add OpenRouter embeddings API support - <a href="https://github.com/BerriAI/litellm/pull/18391" target="_blank" rel="noopener noreferrer">PR #18391</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong>
<ul>
<li>Pass server_tool_use and tool_search_tool_result blocks - <a href="https://github.com/BerriAI/litellm/pull/18770" target="_blank" rel="noopener noreferrer">PR #18770</a></li>
<li>Add Anthropic cache control option to image tool call results - <a href="https://github.com/BerriAI/litellm/pull/18674" target="_blank" rel="noopener noreferrer">PR #18674</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/ollama">Ollama</a></strong>
<ul>
<li>Add dimensions for ollama embedding - <a href="https://github.com/BerriAI/litellm/pull/18536" target="_blank" rel="noopener noreferrer">PR #18536</a></li>
<li>Extract pure base64 data from data URLs for Ollama - <a href="https://github.com/BerriAI/litellm/pull/18465" target="_blank" rel="noopener noreferrer">PR #18465</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/watsonx/index">Watsonx</a></strong>
<ul>
<li>Add Watsonx fields support - <a href="https://github.com/BerriAI/litellm/pull/18569" target="_blank" rel="noopener noreferrer">PR #18569</a></li>
<li>Fix Watsonx Audio Transcription - filter model field - <a href="https://github.com/BerriAI/litellm/pull/18810" target="_blank" rel="noopener noreferrer">PR #18810</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/sap">SAP</a></strong>
<ul>
<li>Add SAP creds for list in proxy UI - <a href="https://github.com/BerriAI/litellm/pull/18375" target="_blank" rel="noopener noreferrer">PR #18375</a></li>
<li>Pass through extra params from allowed_openai_params - <a href="https://github.com/BerriAI/litellm/pull/18432" target="_blank" rel="noopener noreferrer">PR #18432</a></li>
<li>Add client header for SAP AI Core Tracking - <a href="https://github.com/BerriAI/litellm/pull/18714" target="_blank" rel="noopener noreferrer">PR #18714</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/fireworks_ai">Fireworks AI</a></strong>
<ul>
<li>Correct deepseek-v3p2 pricing - <a href="https://github.com/BerriAI/litellm/pull/18483" target="_blank" rel="noopener noreferrer">PR #18483</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/zai">ZAI</a></strong>
<ul>
<li>Add GLM-4.7 model with reasoning support - <a href="https://github.com/BerriAI/litellm/pull/18476" target="_blank" rel="noopener noreferrer">PR #18476</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/codestral">Codestral</a></strong>
<ul>
<li>Correctly route codestral chat and FIM endpoints - <a href="https://github.com/BerriAI/litellm/pull/18467" target="_blank" rel="noopener noreferrer">PR #18467</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/azure_ai">Azure AI</a></strong>
<ul>
<li>Fix authentication errors at messages API via azure_ai - <a href="https://github.com/BerriAI/litellm/pull/18500" target="_blank" rel="noopener noreferrer">PR #18500</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-provider-support">New Provider Support<a href="https://docs.litellm.ai/release_notes/v1-80-15#new-provider-support" class="hash-link" aria-label="Direct link to New Provider Support" title="Direct link to New Provider Support">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/providers/aws_polly">AWS Polly</a></strong> - Add AWS Polly API for TTS - <a href="https://github.com/BerriAI/litellm/pull/18326" target="_blank" rel="noopener noreferrer">PR #18326</a></li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/gigachat">GigaChat</a></strong> - Add GigaChat provider support - <a href="https://github.com/BerriAI/litellm/pull/18564" target="_blank" rel="noopener noreferrer">PR #18564</a></li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/llamagate">LlamaGate</a></strong> - Add LlamaGate as a new provider - <a href="https://github.com/BerriAI/litellm/pull/18673" target="_blank" rel="noopener noreferrer">PR #18673</a></li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/abliteration">Abliteration AI</a></strong> - Add abliteration.ai provider - <a href="https://github.com/BerriAI/litellm/pull/18678" target="_blank" rel="noopener noreferrer">PR #18678</a></li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/manus">Manus</a></strong> - Add Manus API support on /responses, GET /responses - <a href="https://github.com/BerriAI/litellm/pull/18804" target="_blank" rel="noopener noreferrer">PR #18804</a></li>
<li><strong>5 AI Providers via openai_like</strong> - Add 5 AI providers using openai_like - <a href="https://github.com/BerriAI/litellm/pull/18362" target="_blank" rel="noopener noreferrer">PR #18362</a></li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-80-15#bug-fixes" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Gemini</a></strong>
<ul>
<li>Properly catch context window exceeded errors - <a href="https://github.com/BerriAI/litellm/pull/18283" target="_blank" rel="noopener noreferrer">PR #18283</a></li>
<li>Remove prompt caching headers as support has been removed - <a href="https://github.com/BerriAI/litellm/pull/18579" target="_blank" rel="noopener noreferrer">PR #18579</a></li>
<li>Fix generate content request with audio file id - <a href="https://github.com/BerriAI/litellm/pull/18745" target="_blank" rel="noopener noreferrer">PR #18745</a></li>
<li>Fix google_genai streaming adapter provider handling - <a href="https://github.com/BerriAI/litellm/pull/18845" target="_blank" rel="noopener noreferrer">PR #18845</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/groq">Groq</a></strong>
<ul>
<li>Remove deprecated Groq models and update model registry - <a href="https://github.com/BerriAI/litellm/pull/18062" target="_blank" rel="noopener noreferrer">PR #18062</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/vertex">Vertex AI</a></strong>
<ul>
<li>Handle unsupported region for Vertex AI count tokens endpoint - <a href="https://github.com/BerriAI/litellm/pull/18665" target="_blank" rel="noopener noreferrer">PR #18665</a></li>
</ul>
</li>
<li><strong>General</strong>
<ul>
<li>Fix request body for image embedding request - <a href="https://github.com/BerriAI/litellm/pull/18336" target="_blank" rel="noopener noreferrer">PR #18336</a></li>
<li>Fix lost tool_calls when streaming has both text and tool_calls - <a href="https://github.com/BerriAI/litellm/pull/18316" target="_blank" rel="noopener noreferrer">PR #18316</a></li>
<li>Add all resolution for gpt-image-1.5 - <a href="https://github.com/BerriAI/litellm/pull/18586" target="_blank" rel="noopener noreferrer">PR #18586</a></li>
<li>Fix gpt-image-1 cost calculation using token-based pricing - <a href="https://github.com/BerriAI/litellm/pull/17906" target="_blank" rel="noopener noreferrer">PR #17906</a></li>
<li>Fix response_format leaking into extra_body - <a href="https://github.com/BerriAI/litellm/pull/18859" target="_blank" rel="noopener noreferrer">PR #18859</a></li>
<li>Align max_tokens with max_output_tokens for consistency - <a href="https://github.com/BerriAI/litellm/pull/18820" target="_blank" rel="noopener noreferrer">PR #18820</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-80-15#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-80-15#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/response_api">Responses API</a></strong>
<ul>
<li>Add new compact endpoint (v1/responses/compact) - <a href="https://github.com/BerriAI/litellm/pull/18697" target="_blank" rel="noopener noreferrer">PR #18697</a></li>
<li>Support more streaming callback hooks - <a href="https://github.com/BerriAI/litellm/pull/18513" target="_blank" rel="noopener noreferrer">PR #18513</a></li>
<li>Add mapping for reasoning effort to summary param - <a href="https://github.com/BerriAI/litellm/pull/18635" target="_blank" rel="noopener noreferrer">PR #18635</a></li>
<li>Add output_text property to ResponsesAPIResponse - <a href="https://github.com/BerriAI/litellm/pull/18491" target="_blank" rel="noopener noreferrer">PR #18491</a></li>
<li>Add annotations to completions responses API bridge - <a href="https://github.com/BerriAI/litellm/pull/18754" target="_blank" rel="noopener noreferrer">PR #18754</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/interactions">Interactions API</a></strong>
<ul>
<li>Allow using all LiteLLM providers (interactions -&gt; responses API bridge) - <a href="https://github.com/BerriAI/litellm/pull/18373" target="_blank" rel="noopener noreferrer">PR #18373</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/search/index">RAG Search API</a></strong>
<ul>
<li>Add RAG Search/Query endpoint - <a href="https://github.com/BerriAI/litellm/pull/18376" target="_blank" rel="noopener noreferrer">PR #18376</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/anthropic_count_tokens">CountTokens API</a></strong>
<ul>
<li>Add Bedrock as a new provider for <code>/v1/messages/count_tokens</code> - <a href="https://github.com/BerriAI/litellm/pull/18858" target="_blank" rel="noopener noreferrer">PR #18858</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Generate Content</a></strong>
<ul>
<li>Add generate content in LLM route - <a href="https://github.com/BerriAI/litellm/pull/18405" target="_blank" rel="noopener noreferrer">PR #18405</a></li>
</ul>
</li>
<li><strong>General</strong>
<ul>
<li>Enable async_post_call_failure_hook to transform error responses - <a href="https://github.com/BerriAI/litellm/pull/18348" target="_blank" rel="noopener noreferrer">PR #18348</a></li>
<li>Calculate total_tokens manually if missing and can be calculated - <a href="https://github.com/BerriAI/litellm/pull/18445" target="_blank" rel="noopener noreferrer">PR #18445</a></li>
<li>Add custom llm provider to get_llm_provider when sent via UI - <a href="https://github.com/BerriAI/litellm/pull/18638" target="_blank" rel="noopener noreferrer">PR #18638</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs">Bugs<a href="https://docs.litellm.ai/release_notes/v1-80-15#bugs" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>General</strong>
<ul>
<li>Handle empty error objects in response conversion - <a href="https://github.com/BerriAI/litellm/pull/18493" target="_blank" rel="noopener noreferrer">PR #18493</a></li>
<li>Preserve client error status codes in streaming mode - <a href="https://github.com/BerriAI/litellm/pull/18698" target="_blank" rel="noopener noreferrer">PR #18698</a></li>
<li>Return json error response instead of SSE format for initial streaming errors - <a href="https://github.com/BerriAI/litellm/pull/18757" target="_blank" rel="noopener noreferrer">PR #18757</a></li>
<li>Fix auth header for custom api base in generateContent request - <a href="https://github.com/BerriAI/litellm/pull/18637" target="_blank" rel="noopener noreferrer">PR #18637</a></li>
<li>Tool content should be string for Deepinfra - <a href="https://github.com/BerriAI/litellm/pull/18739" target="_blank" rel="noopener noreferrer">PR #18739</a></li>
<li>Fix incomplete usage in response object passed - <a href="https://github.com/BerriAI/litellm/pull/18799" target="_blank" rel="noopener noreferrer">PR #18799</a></li>
<li>Unify model names to provider-defined names - <a href="https://github.com/BerriAI/litellm/pull/18573" target="_blank" rel="noopener noreferrer">PR #18573</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-80-15#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-80-15#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li><strong>SSO Configuration</strong>
<ul>
<li>Add SSO Role Mapping feature - <a href="https://github.com/BerriAI/litellm/pull/18090" target="_blank" rel="noopener noreferrer">PR #18090</a></li>
<li>Add SSO Settings Page - <a href="https://github.com/BerriAI/litellm/pull/18600" target="_blank" rel="noopener noreferrer">PR #18600</a></li>
<li>Allow adding role mappings for SSO - <a href="https://github.com/BerriAI/litellm/pull/18593" target="_blank" rel="noopener noreferrer">PR #18593</a></li>
<li>SSO Settings Page Add Role Mappings - <a href="https://github.com/BerriAI/litellm/pull/18677" target="_blank" rel="noopener noreferrer">PR #18677</a></li>
<li>SSO Settings Loading State + Deprecate Previous SSO Flow - <a href="https://github.com/BerriAI/litellm/pull/18617" target="_blank" rel="noopener noreferrer">PR #18617</a></li>
</ul>
</li>
<li><strong>Virtual Keys</strong>
<ul>
<li>Allow deleting key expiry - <a href="https://github.com/BerriAI/litellm/pull/18278" target="_blank" rel="noopener noreferrer">PR #18278</a></li>
<li>Add optional query param "expand" to /key/list - <a href="https://github.com/BerriAI/litellm/pull/18502" target="_blank" rel="noopener noreferrer">PR #18502</a></li>
<li>Key Table Loading Skeleton - <a href="https://github.com/BerriAI/litellm/pull/18527" target="_blank" rel="noopener noreferrer">PR #18527</a></li>
<li>Allow column resizing on Keys Table - <a href="https://github.com/BerriAI/litellm/pull/18424" target="_blank" rel="noopener noreferrer">PR #18424</a></li>
<li>Virtual Keys Table Loading State Between Pages - <a href="https://github.com/BerriAI/litellm/pull/18619" target="_blank" rel="noopener noreferrer">PR #18619</a></li>
<li>Key and Team Router Setting - <a href="https://github.com/BerriAI/litellm/pull/18790" target="_blank" rel="noopener noreferrer">PR #18790</a></li>
<li>Allow router_settings on Keys and Teams - <a href="https://github.com/BerriAI/litellm/pull/18675" target="_blank" rel="noopener noreferrer">PR #18675</a></li>
<li>Use timedelta to calculate key expiry on generate - <a href="https://github.com/BerriAI/litellm/pull/18666" target="_blank" rel="noopener noreferrer">PR #18666</a></li>
</ul>
</li>
<li><strong>Models + Endpoints</strong>
<ul>
<li>Add Model Clearer Flow For Team Admins - <a href="https://github.com/BerriAI/litellm/pull/18532" target="_blank" rel="noopener noreferrer">PR #18532</a></li>
<li>Model Page Loading State - <a href="https://github.com/BerriAI/litellm/pull/18574" target="_blank" rel="noopener noreferrer">PR #18574</a></li>
<li>Model Page Model Provider Select Performance - <a href="https://github.com/BerriAI/litellm/pull/18425" target="_blank" rel="noopener noreferrer">PR #18425</a></li>
<li>Model Page Sorting Sorts Entire Set - <a href="https://github.com/BerriAI/litellm/pull/18420" target="_blank" rel="noopener noreferrer">PR #18420</a></li>
<li>Refactor Model Hub Page - <a href="https://github.com/BerriAI/litellm/pull/18568" target="_blank" rel="noopener noreferrer">PR #18568</a></li>
<li>Add request provider form on UI - <a href="https://github.com/BerriAI/litellm/pull/18704" target="_blank" rel="noopener noreferrer">PR #18704</a></li>
</ul>
</li>
<li><strong>Organizations &amp; Teams</strong>
<ul>
<li>Allow Organization Admins to See Organization Tab - <a href="https://github.com/BerriAI/litellm/pull/18400" target="_blank" rel="noopener noreferrer">PR #18400</a></li>
<li>Resolve Organization Alias on Team Table - <a href="https://github.com/BerriAI/litellm/pull/18401" target="_blank" rel="noopener noreferrer">PR #18401</a></li>
<li>Resolve Team Alias in Organization Info View - <a href="https://github.com/BerriAI/litellm/pull/18404" target="_blank" rel="noopener noreferrer">PR #18404</a></li>
<li>Allow Organization Admins to View Their Organization Info - <a href="https://github.com/BerriAI/litellm/pull/18417" target="_blank" rel="noopener noreferrer">PR #18417</a></li>
<li>Allow editing team_member_budget_duration in /team/update - <a href="https://github.com/BerriAI/litellm/pull/18735" target="_blank" rel="noopener noreferrer">PR #18735</a></li>
<li>Reusable Duration Select + Team Update Member Budget Duration - <a href="https://github.com/BerriAI/litellm/pull/18736" target="_blank" rel="noopener noreferrer">PR #18736</a></li>
</ul>
</li>
<li><strong>Usage &amp; Spend</strong>
<ul>
<li>Add Error Code Filtering on Spend Logs - <a href="https://github.com/BerriAI/litellm/pull/18359" target="_blank" rel="noopener noreferrer">PR #18359</a></li>
<li>Add Error Code Filtering on UI - <a href="https://github.com/BerriAI/litellm/pull/18366" target="_blank" rel="noopener noreferrer">PR #18366</a></li>
<li>Usage Page User Max Budget fix - <a href="https://github.com/BerriAI/litellm/pull/18555" target="_blank" rel="noopener noreferrer">PR #18555</a></li>
<li>Add endpoint to Daily Activity Tables - <a href="https://github.com/BerriAI/litellm/pull/18729" target="_blank" rel="noopener noreferrer">PR #18729</a></li>
<li>Endpoint Activity in Usage - <a href="https://github.com/BerriAI/litellm/pull/18798" target="_blank" rel="noopener noreferrer">PR #18798</a></li>
</ul>
</li>
<li><strong>Cost Estimator</strong>
<ul>
<li>Add Cost Estimator for AI Gateway - <a href="https://github.com/BerriAI/litellm/pull/18643" target="_blank" rel="noopener noreferrer">PR #18643</a></li>
<li>Add view for estimating costs across requests - <a href="https://github.com/BerriAI/litellm/pull/18645" target="_blank" rel="noopener noreferrer">PR #18645</a></li>
<li>Allow selecting many models for cost estimator - <a href="https://github.com/BerriAI/litellm/pull/18653" target="_blank" rel="noopener noreferrer">PR #18653</a></li>
</ul>
</li>
<li><strong>CloudZero</strong>
<ul>
<li>Improve Create and Delete Path for CloudZero - <a href="https://github.com/BerriAI/litellm/pull/18263" target="_blank" rel="noopener noreferrer">PR #18263</a></li>
<li>Add CloudZero UI Docs - <a href="https://github.com/BerriAI/litellm/pull/18350" target="_blank" rel="noopener noreferrer">PR #18350</a></li>
</ul>
</li>
<li><strong>Playground</strong>
<ul>
<li>Add MCP test support to completions on Playground - <a href="https://github.com/BerriAI/litellm/pull/18440" target="_blank" rel="noopener noreferrer">PR #18440</a></li>
<li>Add selectable MCP servers to the playground - <a href="https://github.com/BerriAI/litellm/pull/18578" target="_blank" rel="noopener noreferrer">PR #18578</a></li>
<li>Add custom proxy base URL support to Playground - <a href="https://github.com/BerriAI/litellm/pull/18661" target="_blank" rel="noopener noreferrer">PR #18661</a></li>
</ul>
</li>
<li><strong>General UI</strong>
<ul>
<li>UI styling improvements and fixes - <a href="https://github.com/BerriAI/litellm/pull/18310" target="_blank" rel="noopener noreferrer">PR #18310</a></li>
<li>Add reusable "New" badge component for feature highlights - <a href="https://github.com/BerriAI/litellm/pull/18537" target="_blank" rel="noopener noreferrer">PR #18537</a></li>
<li>Hide New Badges - <a href="https://github.com/BerriAI/litellm/pull/18547" target="_blank" rel="noopener noreferrer">PR #18547</a></li>
<li>Change Budget page to Have Tabs - <a href="https://github.com/BerriAI/litellm/pull/18576" target="_blank" rel="noopener noreferrer">PR #18576</a></li>
<li>Clicking on Logo Directs to Correct URL - <a href="https://github.com/BerriAI/litellm/pull/18575" target="_blank" rel="noopener noreferrer">PR #18575</a></li>
<li>Add UI support for configuring meta URLs - <a href="https://github.com/BerriAI/litellm/pull/18580" target="_blank" rel="noopener noreferrer">PR #18580</a></li>
<li>Expire Previous UI Session Tokens on Login - <a href="https://github.com/BerriAI/litellm/pull/18557" target="_blank" rel="noopener noreferrer">PR #18557</a></li>
<li>Add license endpoint - <a href="https://github.com/BerriAI/litellm/pull/18311" target="_blank" rel="noopener noreferrer">PR #18311</a></li>
<li>Router Fields Endpoint + React Query for Router Fields - <a href="https://github.com/BerriAI/litellm/pull/18880" target="_blank" rel="noopener noreferrer">PR #18880</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs-1">Bugs<a href="https://docs.litellm.ai/release_notes/v1-80-15#bugs-1" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>UI Fixes</strong>
<ul>
<li>Fix Key Creation MCP Settings Submit Form Unintentionally - <a href="https://github.com/BerriAI/litellm/pull/18355" target="_blank" rel="noopener noreferrer">PR #18355</a></li>
<li>Fix UI Disappears in Development Environments - <a href="https://github.com/BerriAI/litellm/pull/18399" target="_blank" rel="noopener noreferrer">PR #18399</a></li>
<li>Fix Disable Admin UI Flag - <a href="https://github.com/BerriAI/litellm/pull/18397" target="_blank" rel="noopener noreferrer">PR #18397</a></li>
<li>Remove Model Analytics From Model Page - <a href="https://github.com/BerriAI/litellm/pull/18552" target="_blank" rel="noopener noreferrer">PR #18552</a></li>
<li>Useful Links Remove Modal on Adding Links - <a href="https://github.com/BerriAI/litellm/pull/18602" target="_blank" rel="noopener noreferrer">PR #18602</a></li>
<li>SSO Edit Modal Clear Role Mapping Values on Provider Change - <a href="https://github.com/BerriAI/litellm/pull/18680" target="_blank" rel="noopener noreferrer">PR #18680</a></li>
<li>UI Login Case Sensitivity fix - <a href="https://github.com/BerriAI/litellm/pull/18877" target="_blank" rel="noopener noreferrer">PR #18877</a></li>
</ul>
</li>
<li><strong>API Fixes</strong>
<ul>
<li>Fix User Invite &amp; Key Generation Email Notification Logic - <a href="https://github.com/BerriAI/litellm/pull/18524" target="_blank" rel="noopener noreferrer">PR #18524</a></li>
<li>Normalize Proxy Config Callback - <a href="https://github.com/BerriAI/litellm/pull/18775" target="_blank" rel="noopener noreferrer">PR #18775</a></li>
<li>Return empty data array instead of 500 when no models configured - <a href="https://github.com/BerriAI/litellm/pull/18556" target="_blank" rel="noopener noreferrer">PR #18556</a></li>
<li>Enforce org level max budget - <a href="https://github.com/BerriAI/litellm/pull/18813" target="_blank" rel="noopener noreferrer">PR #18813</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="ai-integrations">AI Integrations<a href="https://docs.litellm.ai/release_notes/v1-80-15#ai-integrations" class="hash-link" aria-label="Direct link to AI Integrations" title="Direct link to AI Integrations">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="new-integrations-4-new-integrations">New Integrations (4 new integrations)<a href="https://docs.litellm.ai/release_notes/v1-80-15#new-integrations-4-new-integrations" class="hash-link" aria-label="Direct link to New Integrations (4 new integrations)" title="Direct link to New Integrations (4 new integrations)">​</a></h3>
<table><thead><tr><th>Integration</th><th>Type</th><th>Description</th></tr></thead><tbody><tr><td><a href="https://docs.litellm.ai/docs/observability/focus">Focus</a></td><td>Logging</td><td>Focus export support for observability - <a href="https://github.com/BerriAI/litellm/pull/18802" target="_blank" rel="noopener noreferrer">PR #18802</a></td></tr><tr><td><a href="https://docs.litellm.ai/docs/observability/signoz">SigNoz</a></td><td>Logging</td><td>SigNoz integration for observability - <a href="https://github.com/BerriAI/litellm/pull/18726" target="_blank" rel="noopener noreferrer">PR #18726</a></td></tr><tr><td><a href="https://docs.litellm.ai/docs/proxy/guardrails/qualifire">Qualifire</a></td><td>Guardrails</td><td>Qualifire guardrails and eval webhook - <a href="https://github.com/BerriAI/litellm/pull/18594" target="_blank" rel="noopener noreferrer">PR #18594</a></td></tr><tr><td><a href="https://docs.litellm.ai/docs/observability/levo_integration">Levo AI</a></td><td>Guardrails</td><td>Levo AI integration for security - <a href="https://github.com/BerriAI/litellm/pull/18529" target="_blank" rel="noopener noreferrer">PR #18529</a></td></tr></tbody></table>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="logging">Logging<a href="https://docs.litellm.ai/release_notes/v1-80-15#logging" class="hash-link" aria-label="Direct link to Logging" title="Direct link to Logging">​</a></h3>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging#datadog">DataDog</a></strong>
<ul>
<li>Fix span kind fallback when parent_id missing - <a href="https://github.com/BerriAI/litellm/pull/18418" target="_blank" rel="noopener noreferrer">PR #18418</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging#langfuse">Langfuse</a></strong>
<ul>
<li>Map Gemini cached_tokens to Langfuse cache_read_input_tokens - <a href="https://github.com/BerriAI/litellm/pull/18614" target="_blank" rel="noopener noreferrer">PR #18614</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging#prometheus">Prometheus</a></strong>
<ul>
<li>Align prometheus metric names with DEFINED_PROMETHEUS_METRICS - <a href="https://github.com/BerriAI/litellm/pull/18463" target="_blank" rel="noopener noreferrer">PR #18463</a></li>
<li>Add Prometheus metrics for request queue time and guardrails - <a href="https://github.com/BerriAI/litellm/pull/17973" target="_blank" rel="noopener noreferrer">PR #17973</a></li>
<li>Add caching metrics for cache hits, misses, and tokens - <a href="https://github.com/BerriAI/litellm/pull/18755" target="_blank" rel="noopener noreferrer">PR #18755</a></li>
<li>Skip metrics for invalid API key requests - <a href="https://github.com/BerriAI/litellm/pull/18788" target="_blank" rel="noopener noreferrer">PR #18788</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging#braintrust">Braintrust</a></strong>
<ul>
<li>Pass span_attributes in async logging and skip tags on non-root spans - <a href="https://github.com/BerriAI/litellm/pull/18409" target="_blank" rel="noopener noreferrer">PR #18409</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging#cloudzero">CloudZero</a></strong>
<ul>
<li>Add user email to CloudZero - <a href="https://github.com/BerriAI/litellm/pull/18584" target="_blank" rel="noopener noreferrer">PR #18584</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging#opentelemetry">OpenTelemetry</a></strong>
<ul>
<li>Use already configured opentelemetry providers - <a href="https://github.com/BerriAI/litellm/pull/18279" target="_blank" rel="noopener noreferrer">PR #18279</a></li>
<li>Prevent LiteLLM from closing external OTEL spans - <a href="https://github.com/BerriAI/litellm/pull/18553" target="_blank" rel="noopener noreferrer">PR #18553</a></li>
<li>Allow configuring arize project name for OpenTelemetry service name - <a href="https://github.com/BerriAI/litellm/pull/18738" target="_blank" rel="noopener noreferrer">PR #18738</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging#langsmith">LangSmith</a></strong>
<ul>
<li>Add support for LangSmith organization-scoped API keys with tenant ID - <a href="https://github.com/BerriAI/litellm/pull/18623" target="_blank" rel="noopener noreferrer">PR #18623</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging#generic-api-logger">Generic API Logger</a></strong>
<ul>
<li>Add log_format option to GenericAPILogger - <a href="https://github.com/BerriAI/litellm/pull/18587" target="_blank" rel="noopener noreferrer">PR #18587</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails">Guardrails<a href="https://docs.litellm.ai/release_notes/v1-80-15#guardrails" class="hash-link" aria-label="Direct link to Guardrails" title="Direct link to Guardrails">​</a></h3>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/litellm_content_filter">Content Filter</a></strong>
<ul>
<li>Add content filter logs page - <a href="https://github.com/BerriAI/litellm/pull/18335" target="_blank" rel="noopener noreferrer">PR #18335</a></li>
<li>Log actual event type for guardrails - <a href="https://github.com/BerriAI/litellm/pull/18489" target="_blank" rel="noopener noreferrer">PR #18489</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/qualifire">Qualifire</a></strong>
<ul>
<li>Add Qualifire eval webhook - <a href="https://github.com/BerriAI/litellm/pull/18836" target="_blank" rel="noopener noreferrer">PR #18836</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/lasso_security">Lasso Security</a></strong>
<ul>
<li>Add Lasso guardrail API docs - <a href="https://github.com/BerriAI/litellm/pull/18652" target="_blank" rel="noopener noreferrer">PR #18652</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/noma_security">Noma Security</a></strong>
<ul>
<li>Add MCP guardrail support for Noma - <a href="https://github.com/BerriAI/litellm/pull/18668" target="_blank" rel="noopener noreferrer">PR #18668</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/bedrock">Bedrock Guardrails</a></strong>
<ul>
<li>Remove redundant Bedrock guardrail block handling - <a href="https://github.com/BerriAI/litellm/pull/18634" target="_blank" rel="noopener noreferrer">PR #18634</a></li>
</ul>
</li>
<li><strong>General</strong>
<ul>
<li>Generic guardrail API update - <a href="https://github.com/BerriAI/litellm/pull/18647" target="_blank" rel="noopener noreferrer">PR #18647</a></li>
<li>Prevent proxy startup failures from case-sensitive tool permission guardrail validation - <a href="https://github.com/BerriAI/litellm/pull/18662" target="_blank" rel="noopener noreferrer">PR #18662</a></li>
<li>Extend case normalization to ALL guardrail types - <a href="https://github.com/BerriAI/litellm/pull/18664" target="_blank" rel="noopener noreferrer">PR #18664</a></li>
<li>Fix MCP handling in unified guardrail - <a href="https://github.com/BerriAI/litellm/pull/18630" target="_blank" rel="noopener noreferrer">PR #18630</a></li>
<li>Fix embeddings calltype for guardrail precallhook - <a href="https://github.com/BerriAI/litellm/pull/18740" target="_blank" rel="noopener noreferrer">PR #18740</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-80-15#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li><strong>Platform Fee / Margins</strong> - Add support for Platform Fee / Margins - <a href="https://github.com/BerriAI/litellm/pull/18427" target="_blank" rel="noopener noreferrer">PR #18427</a></li>
<li><strong>Negative Budget Validation</strong> - Add validation for negative budget - <a href="https://github.com/BerriAI/litellm/pull/18583" target="_blank" rel="noopener noreferrer">PR #18583</a></li>
<li><strong>Cost Calculation Fixes</strong>
<ul>
<li>Correct cost calculation when reasoning_tokens are without text_tokens - <a href="https://github.com/BerriAI/litellm/pull/18607" target="_blank" rel="noopener noreferrer">PR #18607</a></li>
<li>Fix background cost tracking tests - <a href="https://github.com/BerriAI/litellm/pull/18588" target="_blank" rel="noopener noreferrer">PR #18588</a></li>
</ul>
</li>
<li><strong>Tag Routing</strong> - Support toggling tag matching between ANY and ALL - <a href="https://github.com/BerriAI/litellm/pull/18776" target="_blank" rel="noopener noreferrer">PR #18776</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-gateway">MCP Gateway<a href="https://docs.litellm.ai/release_notes/v1-80-15#mcp-gateway" class="hash-link" aria-label="Direct link to MCP Gateway" title="Direct link to MCP Gateway">​</a></h2>
<ul>
<li><strong>MCP Global Mode</strong> - Add MCP global mode - <a href="https://github.com/BerriAI/litellm/pull/18639" target="_blank" rel="noopener noreferrer">PR #18639</a></li>
<li><strong>MCP Server Visibility</strong> - Add configurable MCP server visibility - <a href="https://github.com/BerriAI/litellm/pull/18681" target="_blank" rel="noopener noreferrer">PR #18681</a></li>
<li><strong>MCP Registry</strong> - Add MCP registry - <a href="https://github.com/BerriAI/litellm/pull/18850" target="_blank" rel="noopener noreferrer">PR #18850</a></li>
<li><strong>MCP Stdio Header</strong> - Support MCP stdio header env overrides - <a href="https://github.com/BerriAI/litellm/pull/18324" target="_blank" rel="noopener noreferrer">PR #18324</a></li>
<li><strong>Parallel Tool Fetching</strong> - Parallelize tool fetching from multiple MCP servers - <a href="https://github.com/BerriAI/litellm/pull/18627" target="_blank" rel="noopener noreferrer">PR #18627</a></li>
<li><strong>Optimize MCP Server Listing</strong> - Separate health checks for optimized listing - <a href="https://github.com/BerriAI/litellm/pull/18530" target="_blank" rel="noopener noreferrer">PR #18530</a></li>
<li><strong>Auth Improvements</strong>
<ul>
<li>Require auth for MCP connection test endpoint - <a href="https://github.com/BerriAI/litellm/pull/18290" target="_blank" rel="noopener noreferrer">PR #18290</a></li>
<li>Fix MCP gateway OAuth2 auth issues and ClosedResourceError - <a href="https://github.com/BerriAI/litellm/pull/18281" target="_blank" rel="noopener noreferrer">PR #18281</a></li>
</ul>
</li>
<li><strong>Bug Fixes</strong>
<ul>
<li>Fix MCP server health status reporting - <a href="https://github.com/BerriAI/litellm/pull/18443" target="_blank" rel="noopener noreferrer">PR #18443</a></li>
<li>Fix OpenAPI to MCP tool conversion - <a href="https://github.com/BerriAI/litellm/pull/18597" target="_blank" rel="noopener noreferrer">PR #18597</a></li>
<li>Remove exec() usage and handle invalid OpenAPI parameter names for security - <a href="https://github.com/BerriAI/litellm/pull/18480" target="_blank" rel="noopener noreferrer">PR #18480</a></li>
<li>Fix MCP error when using multiple servers simultaneously - <a href="https://github.com/BerriAI/litellm/pull/18855" target="_blank" rel="noopener noreferrer">PR #18855</a></li>
</ul>
</li>
<li><strong>Migrate MCP Fetching Logic to React Query</strong> - <a href="https://github.com/BerriAI/litellm/pull/18352" target="_blank" rel="noopener noreferrer">PR #18352</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements">Performance / Loadbalancing / Reliability improvements<a href="https://docs.litellm.ai/release_notes/v1-80-15#performance--loadbalancing--reliability-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements" title="Direct link to Performance / Loadbalancing / Reliability improvements">​</a></h2>
<ul>
<li><strong>92.7% Faster Provider Config Lookup</strong> - LiteLLM now stresses LLM providers 2.5x more - <a href="https://github.com/BerriAI/litellm/pull/18867" target="_blank" rel="noopener noreferrer">PR #18867</a></li>
<li><strong>Lazy Loading Improvements</strong>
<ul>
<li>Consolidate lazy import handlers with registry pattern - <a href="https://github.com/BerriAI/litellm/pull/18389" target="_blank" rel="noopener noreferrer">PR #18389</a></li>
<li>Complete lazy loading migration for all 180+ LLM config classes - <a href="https://github.com/BerriAI/litellm/pull/18392" target="_blank" rel="noopener noreferrer">PR #18392</a></li>
<li>Lazy load additional components (types, callbacks, utilities) - <a href="https://github.com/BerriAI/litellm/pull/18396" target="_blank" rel="noopener noreferrer">PR #18396</a></li>
<li>Add lazy loading for get_llm_provider - <a href="https://github.com/BerriAI/litellm/pull/18591" target="_blank" rel="noopener noreferrer">PR #18591</a></li>
<li>Lazy-load heavy audio library and loggers - <a href="https://github.com/BerriAI/litellm/pull/18592" target="_blank" rel="noopener noreferrer">PR #18592</a></li>
<li>Lazy load 9 heavy imports in litellm/utils.py - <a href="https://github.com/BerriAI/litellm/pull/18595" target="_blank" rel="noopener noreferrer">PR #18595</a></li>
<li>Lazy load heavy imports to improve import time and memory usage - <a href="https://github.com/BerriAI/litellm/pull/18610" target="_blank" rel="noopener noreferrer">PR #18610</a></li>
<li>Implement lazy loading for provider configs, model info classes, streaming handlers - <a href="https://github.com/BerriAI/litellm/pull/18611" target="_blank" rel="noopener noreferrer">PR #18611</a></li>
<li>Lazy load 15 additional imports - <a href="https://github.com/BerriAI/litellm/pull/18613" target="_blank" rel="noopener noreferrer">PR #18613</a></li>
<li>Lazy load 15+ unused imports - <a href="https://github.com/BerriAI/litellm/pull/18616" target="_blank" rel="noopener noreferrer">PR #18616</a></li>
<li>Lazy load DatadogLLMObsInitParams - <a href="https://github.com/BerriAI/litellm/pull/18658" target="_blank" rel="noopener noreferrer">PR #18658</a></li>
<li>Migrate utils.py lazy imports to registry pattern - <a href="https://github.com/BerriAI/litellm/pull/18657" target="_blank" rel="noopener noreferrer">PR #18657</a></li>
<li>Lazy load get_llm_provider and remove_index_from_tool_calls - <a href="https://github.com/BerriAI/litellm/pull/18608" target="_blank" rel="noopener noreferrer">PR #18608</a></li>
</ul>
</li>
<li><strong>Router Improvements</strong>
<ul>
<li>Validate routing_strategy at startup to fail fast with helpful error - <a href="https://github.com/BerriAI/litellm/pull/18624" target="_blank" rel="noopener noreferrer">PR #18624</a></li>
<li>Correct num_retries tracking in retry logic - <a href="https://github.com/BerriAI/litellm/pull/18712" target="_blank" rel="noopener noreferrer">PR #18712</a></li>
<li>Improve error messages and validation for wildcard routing with multiple credentials - <a href="https://github.com/BerriAI/litellm/pull/18629" target="_blank" rel="noopener noreferrer">PR #18629</a></li>
</ul>
</li>
<li><strong>Memory Improvements</strong>
<ul>
<li>Add memory pattern detection test and fix bad memory patterns - <a href="https://github.com/BerriAI/litellm/pull/18589" target="_blank" rel="noopener noreferrer">PR #18589</a></li>
<li>Add unbounded data structure detection to memory test - <a href="https://github.com/BerriAI/litellm/pull/18590" target="_blank" rel="noopener noreferrer">PR #18590</a></li>
<li>Add memory leak detection tests with CI integration - <a href="https://github.com/BerriAI/litellm/pull/18881" target="_blank" rel="noopener noreferrer">PR #18881</a></li>
</ul>
</li>
<li><strong>Database</strong>
<ul>
<li>Add idx on LOWER(user_email) for faster duplicate email checks - <a href="https://github.com/BerriAI/litellm/pull/18828" target="_blank" rel="noopener noreferrer">PR #18828</a></li>
<li>Proactive RDS IAM token refresh to prevent 15-min connection failed - <a href="https://github.com/BerriAI/litellm/pull/18795" target="_blank" rel="noopener noreferrer">PR #18795</a></li>
<li>Clarify database_connection_pool_limit applies per worker - <a href="https://github.com/BerriAI/litellm/pull/18780" target="_blank" rel="noopener noreferrer">PR #18780</a></li>
<li>Make base_connection_pool_limit default value the same - <a href="https://github.com/BerriAI/litellm/pull/18721" target="_blank" rel="noopener noreferrer">PR #18721</a></li>
</ul>
</li>
<li><strong>Docker</strong>
<ul>
<li>Add libsndfile to database Docker image for audio processing - <a href="https://github.com/BerriAI/litellm/pull/18612" target="_blank" rel="noopener noreferrer">PR #18612</a></li>
<li>Add line_profiler support for performance analysis and fix Windows CRLF issues - <a href="https://github.com/BerriAI/litellm/pull/18773" target="_blank" rel="noopener noreferrer">PR #18773</a></li>
</ul>
</li>
<li><strong>Helm</strong>
<ul>
<li>Add lifecycle support to Helm charts - <a href="https://github.com/BerriAI/litellm/pull/18517" target="_blank" rel="noopener noreferrer">PR #18517</a></li>
</ul>
</li>
<li><strong>Authentication</strong>
<ul>
<li>Add Kubernetes ServiceAccount JWT authentication support - <a href="https://github.com/BerriAI/litellm/pull/18055" target="_blank" rel="noopener noreferrer">PR #18055</a></li>
<li>Use async anthropic client to prevent event loop blocking - <a href="https://github.com/BerriAI/litellm/pull/18435" target="_blank" rel="noopener noreferrer">PR #18435</a></li>
</ul>
</li>
<li><strong>Logging Worker</strong>
<ul>
<li>Handle event loop changes in multiprocessing - <a href="https://github.com/BerriAI/litellm/pull/18423" target="_blank" rel="noopener noreferrer">PR #18423</a></li>
</ul>
</li>
<li><strong>Security</strong>
<ul>
<li>Prevent expired key plaintext leak in error response - <a href="https://github.com/BerriAI/litellm/pull/18860" target="_blank" rel="noopener noreferrer">PR #18860</a></li>
<li>Mask extra header secrets in model info - <a href="https://github.com/BerriAI/litellm/pull/18822" target="_blank" rel="noopener noreferrer">PR #18822</a></li>
<li>Prevent duplicate User-Agent tags in request_tags - <a href="https://github.com/BerriAI/litellm/pull/18723" target="_blank" rel="noopener noreferrer">PR #18723</a></li>
<li>Properly use litellm api keys - <a href="https://github.com/BerriAI/litellm/pull/18832" target="_blank" rel="noopener noreferrer">PR #18832</a></li>
</ul>
</li>
<li><strong>Misc</strong>
<ul>
<li>Remove double imports in main.py - <a href="https://github.com/BerriAI/litellm/pull/18406" target="_blank" rel="noopener noreferrer">PR #18406</a></li>
<li>Add LITELLM_DISABLE_LAZY_LOADING env var to fix VCR cassette creation issue - <a href="https://github.com/BerriAI/litellm/pull/18725" target="_blank" rel="noopener noreferrer">PR #18725</a></li>
<li>Add xiaomi_mimo to LlmProviders enum to fix router support - <a href="https://github.com/BerriAI/litellm/pull/18819" target="_blank" rel="noopener noreferrer">PR #18819</a></li>
<li>Allow installation with current grpcio on old Python - <a href="https://github.com/BerriAI/litellm/pull/18473" target="_blank" rel="noopener noreferrer">PR #18473</a></li>
<li>Add Custom CA certificates to boto3 clients - <a href="https://github.com/BerriAI/litellm/pull/18852" target="_blank" rel="noopener noreferrer">PR #18852</a></li>
<li>Fix bedrock_cache, metadata and max_model_budget - <a href="https://github.com/BerriAI/litellm/pull/18872" target="_blank" rel="noopener noreferrer">PR #18872</a></li>
<li>Fix LiteLLM SDK embedding headers missing field - <a href="https://github.com/BerriAI/litellm/pull/18844" target="_blank" rel="noopener noreferrer">PR #18844</a></li>
<li>Put automatic reasoning summary inclusion behind feat flag - <a href="https://github.com/BerriAI/litellm/pull/18688" target="_blank" rel="noopener noreferrer">PR #18688</a></li>
<li>turn_off_message_logging Does Not Redact Request Messages in proxy_server_request Field - <a href="https://github.com/BerriAI/litellm/pull/18897" target="_blank" rel="noopener noreferrer">PR #18897</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates">Documentation Updates<a href="https://docs.litellm.ai/release_notes/v1-80-15#documentation-updates" class="hash-link" aria-label="Direct link to Documentation Updates" title="Direct link to Documentation Updates">​</a></h2>
<ul>
<li><strong>Provider Documentation</strong>
<ul>
<li>Update MiniMax docs to be in proper format - <a href="https://github.com/BerriAI/litellm/pull/18403" target="_blank" rel="noopener noreferrer">PR #18403</a></li>
<li>Add docs for 5 AI providers - <a href="https://github.com/BerriAI/litellm/pull/18388" target="_blank" rel="noopener noreferrer">PR #18388</a></li>
<li>Fix gpt-5-mini reasoning_effort supported values - <a href="https://github.com/BerriAI/litellm/pull/18346" target="_blank" rel="noopener noreferrer">PR #18346</a></li>
<li>Fix PDF documentation inconsistency in Anthropic page - <a href="https://github.com/BerriAI/litellm/pull/18816" target="_blank" rel="noopener noreferrer">PR #18816</a></li>
<li>Update OpenRouter docs to include embedding support - <a href="https://github.com/BerriAI/litellm/pull/18874" target="_blank" rel="noopener noreferrer">PR #18874</a></li>
<li>Add LITELLM_REASONING_AUTO_SUMMARY in doc - <a href="https://github.com/BerriAI/litellm/pull/18705" target="_blank" rel="noopener noreferrer">PR #18705</a></li>
</ul>
</li>
<li><strong>MCP Documentation</strong>
<ul>
<li>Agentcore MCP server docs - <a href="https://github.com/BerriAI/litellm/pull/18603" target="_blank" rel="noopener noreferrer">PR #18603</a></li>
<li>Mention MCP prompt/resources types in overview - <a href="https://github.com/BerriAI/litellm/pull/18669" target="_blank" rel="noopener noreferrer">PR #18669</a></li>
<li>Add Focus docs - <a href="https://github.com/BerriAI/litellm/pull/18837" target="_blank" rel="noopener noreferrer">PR #18837</a></li>
</ul>
</li>
<li><strong>Guardrails Documentation</strong>
<ul>
<li>Qualifire docs hotfix - <a href="https://github.com/BerriAI/litellm/pull/18724" target="_blank" rel="noopener noreferrer">PR #18724</a></li>
</ul>
</li>
<li><strong>Infrastructure Documentation</strong>
<ul>
<li>IAM Roles Anywhere docs - <a href="https://github.com/BerriAI/litellm/pull/18559" target="_blank" rel="noopener noreferrer">PR #18559</a></li>
<li>Fix formatting in proxy configs documentation - <a href="https://github.com/BerriAI/litellm/pull/18498" target="_blank" rel="noopener noreferrer">PR #18498</a></li>
<li>Fix GCS cache docs missing for proxy mode - <a href="https://github.com/BerriAI/litellm/pull/13328" target="_blank" rel="noopener noreferrer">PR #13328</a></li>
<li>Fix how to execute cloudzero sql - <a href="https://github.com/BerriAI/litellm/pull/18841" target="_blank" rel="noopener noreferrer">PR #18841</a></li>
</ul>
</li>
<li><strong>General</strong>
<ul>
<li>LiteLLM adopters section - <a href="https://github.com/BerriAI/litellm/pull/18605" target="_blank" rel="noopener noreferrer">PR #18605</a></li>
<li>Remove redundant comments about setting litellm.callbacks - <a href="https://github.com/BerriAI/litellm/pull/18711" target="_blank" rel="noopener noreferrer">PR #18711</a></li>
<li>Update header to be markdown bold by removing space - <a href="https://github.com/BerriAI/litellm/pull/18846" target="_blank" rel="noopener noreferrer">PR #18846</a></li>
<li>Manus docs - new provider - <a href="https://github.com/BerriAI/litellm/pull/18817" target="_blank" rel="noopener noreferrer">PR #18817</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-80-15#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@prasadkona made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18349" target="_blank" rel="noopener noreferrer">PR #18349</a></li>
<li>@lucasrothman made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18283" target="_blank" rel="noopener noreferrer">PR #18283</a></li>
<li>@aggeentik made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18317" target="_blank" rel="noopener noreferrer">PR #18317</a></li>
<li>@mihidumh made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18361" target="_blank" rel="noopener noreferrer">PR #18361</a></li>
<li>@Prazeina made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18498" target="_blank" rel="noopener noreferrer">PR #18498</a></li>
<li>@systec-dk made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18500" target="_blank" rel="noopener noreferrer">PR #18500</a></li>
<li>@xuan07t2 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18514" target="_blank" rel="noopener noreferrer">PR #18514</a></li>
<li>@RensDimmendaal made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18190" target="_blank" rel="noopener noreferrer">PR #18190</a></li>
<li>@yurekami made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18483" target="_blank" rel="noopener noreferrer">PR #18483</a></li>
<li>@agertz7 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18556" target="_blank" rel="noopener noreferrer">PR #18556</a></li>
<li>@yudelevi made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18550" target="_blank" rel="noopener noreferrer">PR #18550</a></li>
<li>@smallp made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18536" target="_blank" rel="noopener noreferrer">PR #18536</a></li>
<li>@kevinpauer made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18569" target="_blank" rel="noopener noreferrer">PR #18569</a></li>
<li>@cansakiroglu made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18517" target="_blank" rel="noopener noreferrer">PR #18517</a></li>
<li>@dee-walia20 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18432" target="_blank" rel="noopener noreferrer">PR #18432</a></li>
<li>@luxinfeng made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18477" target="_blank" rel="noopener noreferrer">PR #18477</a></li>
<li>@cantalupo555 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18476" target="_blank" rel="noopener noreferrer">PR #18476</a></li>
<li>@andersk made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18473" target="_blank" rel="noopener noreferrer">PR #18473</a></li>
<li>@majiayu000 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18467" target="_blank" rel="noopener noreferrer">PR #18467</a></li>
<li>@amangupta-20 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18529" target="_blank" rel="noopener noreferrer">PR #18529</a></li>
<li>@hamzaq453 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18480" target="_blank" rel="noopener noreferrer">PR #18480</a></li>
<li>@ktsaou made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18627" target="_blank" rel="noopener noreferrer">PR #18627</a></li>
<li>@FlibbertyGibbitz made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18624" target="_blank" rel="noopener noreferrer">PR #18624</a></li>
<li>@drorIvry made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18594" target="_blank" rel="noopener noreferrer">PR #18594</a></li>
<li>@urainshah made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18524" target="_blank" rel="noopener noreferrer">PR #18524</a></li>
<li>@mangabits made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18279" target="_blank" rel="noopener noreferrer">PR #18279</a></li>
<li>@0717376 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18564" target="_blank" rel="noopener noreferrer">PR #18564</a></li>
<li>@nmgarza5 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17330" target="_blank" rel="noopener noreferrer">PR #17330</a></li>
<li>@wileykestner made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18445" target="_blank" rel="noopener noreferrer">PR #18445</a></li>
<li>@minijeong-log made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14440" target="_blank" rel="noopener noreferrer">PR #14440</a></li>
<li>@Isaac4real made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18710" target="_blank" rel="noopener noreferrer">PR #18710</a></li>
<li>@marukaz made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18711" target="_blank" rel="noopener noreferrer">PR #18711</a></li>
<li>@rohitravirane made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18712" target="_blank" rel="noopener noreferrer">PR #18712</a></li>
<li>@lizzzcai made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18714" target="_blank" rel="noopener noreferrer">PR #18714</a></li>
<li>@hkd987 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18673" target="_blank" rel="noopener noreferrer">PR #18673</a></li>
<li>@Mr-Pepe made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18674" target="_blank" rel="noopener noreferrer">PR #18674</a></li>
<li>@gkarthi-signoz made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18726" target="_blank" rel="noopener noreferrer">PR #18726</a></li>
<li>@Tianduo16 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18723" target="_blank" rel="noopener noreferrer">PR #18723</a></li>
<li>@wilsonjr made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18721" target="_blank" rel="noopener noreferrer">PR #18721</a></li>
<li>@abliteration-ai made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18678" target="_blank" rel="noopener noreferrer">PR #18678</a></li>
<li>@danialkhan02 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18770" target="_blank" rel="noopener noreferrer">PR #18770</a></li>
<li>@ihower made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18409" target="_blank" rel="noopener noreferrer">PR #18409</a></li>
<li>@elkkhan made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18391" target="_blank" rel="noopener noreferrer">PR #18391</a></li>
<li>@runixer made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18435" target="_blank" rel="noopener noreferrer">PR #18435</a></li>
<li>@choby-shun made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18776" target="_blank" rel="noopener noreferrer">PR #18776</a></li>
<li>@jutaz made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18853" target="_blank" rel="noopener noreferrer">PR #18853</a></li>
<li>@sjmatta made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18250" target="_blank" rel="noopener noreferrer">PR #18250</a></li>
<li>@andres-ortizl made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18856" target="_blank" rel="noopener noreferrer">PR #18856</a></li>
<li>@gauthiermartin made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18844" target="_blank" rel="noopener noreferrer">PR #18844</a></li>
<li>@mel2oo made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18845" target="_blank" rel="noopener noreferrer">PR #18845</a></li>
<li>@DominikHallab made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18846" target="_blank" rel="noopener noreferrer">PR #18846</a></li>
<li>@ji-chuan-che made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18540" target="_blank" rel="noopener noreferrer">PR #18540</a></li>
<li>@raghav-stripe made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18858" target="_blank" rel="noopener noreferrer">PR #18858</a></li>
<li>@akraines made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18629" target="_blank" rel="noopener noreferrer">PR #18629</a></li>
<li>@otaviofbrito made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18665" target="_blank" rel="noopener noreferrer">PR #18665</a></li>
<li>@chetanchoudhary-sumo made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18587" target="_blank" rel="noopener noreferrer">PR #18587</a></li>
<li>@pascalwhoop made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/13328" target="_blank" rel="noopener noreferrer">PR #13328</a></li>
<li>@orgersh92 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18652" target="_blank" rel="noopener noreferrer">PR #18652</a></li>
<li>@DevajMody made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18497" target="_blank" rel="noopener noreferrer">PR #18497</a></li>
<li>@matt-greathouse made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18247" target="_blank" rel="noopener noreferrer">PR #18247</a></li>
<li>@emerzon made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18290" target="_blank" rel="noopener noreferrer">PR #18290</a></li>
<li>@Eric84626 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18281" target="_blank" rel="noopener noreferrer">PR #18281</a></li>
<li>@LukasdeBoer made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18055" target="_blank" rel="noopener noreferrer">PR #18055</a></li>
<li>@LingXuanYin made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18513" target="_blank" rel="noopener noreferrer">PR #18513</a></li>
<li>@krisxia0506 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18698" target="_blank" rel="noopener noreferrer">PR #18698</a></li>
<li>@LouisShark made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18414" target="_blank" rel="noopener noreferrer">PR #18414</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog">Full Changelog<a href="https://docs.litellm.ai/release_notes/v1-80-15#full-changelog" class="hash-link" aria-label="Direct link to Full Changelog" title="Direct link to Full Changelog">​</a></h2>
<p><strong><a href="https://github.com/BerriAI/litellm/compare/v1.80.11.rc.1...v1.80.15-stable.1" target="_blank" rel="noopener noreferrer">View complete changelog on GitHub</a></strong></p>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[v1.80.11-stable - Google Interactions API]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-80-11</id>
        <link href="https://docs.litellm.ai/release_notes/v1-80-11"/>
        <updated>2025-12-20T10:00:00.000Z</updated>
        <summary type="html"><![CDATA[Deploy this version]]></summary>
        <content type="html"><![CDATA[<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-80-11#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">docker.litellm.ai/berriai/litellm:v1.80.11-stable</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.80.11</span><br></span></code></pre></div></div></div></div></div>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-80-11#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<ul>
<li><strong>Gemini 3 Flash Preview</strong> - <a href="https://docs.litellm.ai/docs/providers/gemini">Day 0 support for Google's Gemini 3 Flash Preview with reasoning capabilities</a></li>
<li><strong>Stability AI Image Generation</strong> - <a href="https://docs.litellm.ai/docs/providers/stability">New provider for Stability AI image generation and editing</a></li>
<li><strong>LiteLLM Content Filter</strong> - <a href="https://docs.litellm.ai/docs/proxy/guardrails/litellm_content_filter">Built-in guardrails for harmful content, bias, and PII detection with image support</a></li>
<li><strong>New Provider: Venice.ai</strong> - Support for Venice.ai API via providers.json</li>
<li><strong>Unified Skills API</strong> - <a href="https://docs.litellm.ai/docs/skills">Skills API works across Anthropic, Vertex, Azure, and Bedrock</a></li>
<li><strong>Azure Sentinel Logging</strong> - <a href="https://docs.litellm.ai/docs/observability/azure_sentinel">New logging integration for Azure Sentinel</a></li>
<li><strong>Guardrails Load Balancing</strong> - <a href="https://docs.litellm.ai/docs/proxy/guardrails">Load balance between multiple guardrail providers</a></li>
<li><strong>Email Budget Alerts</strong> - <a href="https://docs.litellm.ai/docs/proxy/email">Send email notifications when budgets are reached</a></li>
<li><strong>Cloudzero Integration on UI</strong> - Setup your Cloudzero Integration Directly on the UI</li>
</ul>
<hr>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="cloudzero-integration-on-ui">Cloudzero Integration on UI<a href="https://docs.litellm.ai/release_notes/v1-80-11#cloudzero-integration-on-ui" class="hash-link" aria-label="Direct link to Cloudzero Integration on UI" title="Direct link to Cloudzero Integration on UI">​</a></h3>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAFCAYAAAB8ZH1oAAAACXBIWXMAAAsTAAALEwEAmpwYAAAAcElEQVR4nG3MOw7CQBRD0dn/5ujpkGhAgsAAk7yfLwoFIhJu7OLIzd0Zc/B8JuZGROARVAlJKBOVaJnJNF3p/cYcry9cOzKpZYFMmlSMMVifVUGGfXZJ/Kat94/eMXMOl+J034INdDcWc3ZH2J//Ot6hi8Rx8IBO1AAAAABJRU5ErkJggg==&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="309"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/ui_cloudzero.8773b85.640.png" srcset="/assets/ideal-img/ui_cloudzero.8773b85.640.png 640w,/assets/ideal-img/ui_cloudzero.7c5cc2f.1005.png 1005w" width="640" height="309"></noscript></div>
<p>Users can now configure their Cloudzero Integration directly on the UI.</p>
<hr>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="performance-50-reduction-in-memory-usage-and-import-latency-for-the-litellm-sdk">Performance: 50% Reduction in Memory Usage and Import Latency for the LiteLLM SDK<a href="https://docs.litellm.ai/release_notes/v1-80-11#performance-50-reduction-in-memory-usage-and-import-latency-for-the-litellm-sdk" class="hash-link" aria-label="Direct link to Performance: 50% Reduction in Memory Usage and Import Latency for the LiteLLM SDK" title="Direct link to Performance: 50% Reduction in Memory Usage and Import Latency for the LiteLLM SDK">​</a></h3>
<p>We've completely restructured <code>litellm.__init__.py</code> to defer heavy imports until they're actually needed, implementing lazy loading for <strong>109 components</strong>.</p>
<p>This refactoring includes <strong>41 provider config classes</strong>, <strong>40 utility functions</strong>, cache implementations (Redis, DualCache, InMemoryCache), HTTP handlers, logging, types, and other heavy dependencies. Heavy libraries like tiktoken and boto3 are now loaded on-demand rather than eagerly at import time.</p>
<p>This makes LiteLLM especially beneficial for serverless functions, Lambda deployments, and containerized environments where cold start times and memory footprint matter.</p>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-providers-and-endpoints">New Providers and Endpoints<a href="https://docs.litellm.ai/release_notes/v1-80-11#new-providers-and-endpoints" class="hash-link" aria-label="Direct link to New Providers and Endpoints" title="Direct link to New Providers and Endpoints">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="new-providers-5-new-providers">New Providers (5 new providers)<a href="https://docs.litellm.ai/release_notes/v1-80-11#new-providers-5-new-providers" class="hash-link" aria-label="Direct link to New Providers (5 new providers)" title="Direct link to New Providers (5 new providers)">​</a></h3>
<table><thead><tr><th>Provider</th><th>Supported LiteLLM Endpoints</th><th>Description</th></tr></thead><tbody><tr><td><a href="https://docs.litellm.ai/docs/providers/stability">Stability AI</a></td><td><code>/images/generations</code>, <code>/images/edits</code></td><td>Stable Diffusion 3, SD3.5, image editing and generation</td></tr><tr><td>Venice.ai</td><td><code>/chat/completions</code>, <code>/messages</code>, <code>/responses</code></td><td>Venice.ai API integration via providers.json</td></tr><tr><td><a href="https://docs.litellm.ai/docs/providers/pydantic_ai_agent">Pydantic AI Agents</a></td><td><code>/a2a</code></td><td>Pydantic AI agents for A2A protocol workflows</td></tr><tr><td><a href="https://docs.litellm.ai/docs/providers/vertex_ai_agent_engine">VertexAI Agent Engine</a></td><td><code>/a2a</code></td><td>Google Vertex AI Agent Engine for agentic workflows</td></tr><tr><td><a href="https://docs.litellm.ai/docs/search/linkup">LinkUp Search</a></td><td><code>/search</code></td><td>LinkUp web search API integration</td></tr></tbody></table>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="new-llm-api-endpoints-2-new-endpoints">New LLM API Endpoints (2 new endpoints)<a href="https://docs.litellm.ai/release_notes/v1-80-11#new-llm-api-endpoints-2-new-endpoints" class="hash-link" aria-label="Direct link to New LLM API Endpoints (2 new endpoints)" title="Direct link to New LLM API Endpoints (2 new endpoints)">​</a></h3>
<table><thead><tr><th>Endpoint</th><th>Method</th><th>Description</th><th>Documentation</th></tr></thead><tbody><tr><td><code>/interactions</code></td><td>POST</td><td>Google Interactions API for conversational AI</td><td><a href="https://docs.litellm.ai/docs/interactions">Docs</a></td></tr><tr><td><code>/search</code></td><td>POST</td><td>RAG Search API with rerankers</td><td><a href="https://docs.litellm.ai/docs/search/index">Docs</a></td></tr></tbody></table>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-80-11#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support-55-new-models">New Model Support (55+ new models)<a href="https://docs.litellm.ai/release_notes/v1-80-11#new-model-support-55-new-models" class="hash-link" aria-label="Direct link to New Model Support (55+ new models)" title="Direct link to New Model Support (55+ new models)">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th><th>Features</th></tr></thead><tbody><tr><td>Gemini</td><td><code>gemini/gemini-3-flash-preview</code></td><td>1M</td><td>$0.50</td><td>$3.00</td><td>Reasoning, vision, audio, video, PDF</td></tr><tr><td>Vertex AI</td><td><code>vertex_ai/gemini-3-flash-preview</code></td><td>1M</td><td>$0.50</td><td>$3.00</td><td>Reasoning, vision, audio, video, PDF</td></tr><tr><td>Azure AI</td><td><code>azure_ai/deepseek-v3.2</code></td><td>164K</td><td>$0.58</td><td>$1.68</td><td>Reasoning, function calling, caching</td></tr><tr><td>Azure AI</td><td><code>azure_ai/cohere-rerank-v4.0-pro</code></td><td>32K</td><td>$0.0025/query</td><td>-</td><td>Rerank</td></tr><tr><td>Azure AI</td><td><code>azure_ai/cohere-rerank-v4.0-fast</code></td><td>32K</td><td>$0.002/query</td><td>-</td><td>Rerank</td></tr><tr><td>OpenRouter</td><td><code>openrouter/openai/gpt-5.2</code></td><td>400K</td><td>$1.75</td><td>$14.00</td><td>Reasoning, vision, caching</td></tr><tr><td>OpenRouter</td><td><code>openrouter/openai/gpt-5.2-pro</code></td><td>400K</td><td>$21.00</td><td>$168.00</td><td>Reasoning, vision</td></tr><tr><td>OpenRouter</td><td><code>openrouter/mistralai/devstral-2512</code></td><td>262K</td><td>$0.15</td><td>$0.60</td><td>Function calling</td></tr><tr><td>OpenRouter</td><td><code>openrouter/mistralai/ministral-3b-2512</code></td><td>131K</td><td>$0.10</td><td>$0.10</td><td>Function calling, vision</td></tr><tr><td>OpenRouter</td><td><code>openrouter/mistralai/ministral-8b-2512</code></td><td>262K</td><td>$0.15</td><td>$0.15</td><td>Function calling, vision</td></tr><tr><td>OpenRouter</td><td><code>openrouter/mistralai/ministral-14b-2512</code></td><td>262K</td><td>$0.20</td><td>$0.20</td><td>Function calling, vision</td></tr><tr><td>OpenRouter</td><td><code>openrouter/mistralai/mistral-large-2512</code></td><td>262K</td><td>$0.50</td><td>$1.50</td><td>Function calling, vision</td></tr><tr><td>OpenAI</td><td><code>gpt-4o-transcribe-diarize</code></td><td>16K</td><td>$6.00/audio</td><td>-</td><td>Audio transcription with diarization</td></tr><tr><td>OpenAI</td><td><code>gpt-image-1.5-2025-12-16</code></td><td>-</td><td>Various</td><td>Various</td><td>Image generation</td></tr><tr><td>Stability</td><td><code>stability/sd3-large</code></td><td>-</td><td>-</td><td>$0.065/image</td><td>Image generation</td></tr><tr><td>Stability</td><td><code>stability/sd3.5-large</code></td><td>-</td><td>-</td><td>$0.065/image</td><td>Image generation</td></tr><tr><td>Stability</td><td><code>stability/stable-image-ultra</code></td><td>-</td><td>-</td><td>$0.08/image</td><td>Image generation</td></tr><tr><td>Stability</td><td><code>stability/inpaint</code></td><td>-</td><td>-</td><td>$0.005/image</td><td>Image editing</td></tr><tr><td>Stability</td><td><code>stability/outpaint</code></td><td>-</td><td>-</td><td>$0.004/image</td><td>Image editing</td></tr><tr><td>Bedrock</td><td><code>stability.stable-conservative-upscale-v1:0</code></td><td>-</td><td>-</td><td>$0.40/image</td><td>Image upscaling</td></tr><tr><td>Bedrock</td><td><code>stability.stable-creative-upscale-v1:0</code></td><td>-</td><td>-</td><td>$0.60/image</td><td>Image upscaling</td></tr><tr><td>Vertex AI</td><td><code>vertex_ai/deepseek-ai/deepseek-ocr-maas</code></td><td>-</td><td>$0.30</td><td>$1.20</td><td>OCR</td></tr><tr><td>LinkUp</td><td><code>linkup/search</code></td><td>-</td><td>$5.87/1K queries</td><td>-</td><td>Web search</td></tr><tr><td>LinkUp</td><td><code>linkup/search-deep</code></td><td>-</td><td>$58.67/1K queries</td><td>-</td><td>Deep web search</td></tr><tr><td>GitHub Copilot</td><td>20+ models</td><td>Various</td><td>-</td><td>-</td><td>Chat completions</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-80-11#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Gemini</a></strong>
<ul>
<li>Add Gemini 3 Flash Preview day 0 support with reasoning - <a href="https://github.com/BerriAI/litellm/pull/18135" target="_blank" rel="noopener noreferrer">PR #18135</a></li>
<li>Support extra_headers in batch embeddings - <a href="https://github.com/BerriAI/litellm/pull/18004" target="_blank" rel="noopener noreferrer">PR #18004</a></li>
<li>Propagate token usage when generating images - <a href="https://github.com/BerriAI/litellm/pull/17987" target="_blank" rel="noopener noreferrer">PR #17987</a></li>
<li>Use JSON instead of form-data for image edit requests - <a href="https://github.com/BerriAI/litellm/pull/18012" target="_blank" rel="noopener noreferrer">PR #18012</a></li>
<li>Fix web search requests count - <a href="https://github.com/BerriAI/litellm/pull/17921" target="_blank" rel="noopener noreferrer">PR #17921</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong>
<ul>
<li>Use dynamic max_tokens based on model - <a href="https://github.com/BerriAI/litellm/pull/17900" target="_blank" rel="noopener noreferrer">PR #17900</a></li>
<li>Fix claude-3-7-sonnet max_tokens to 64K default - <a href="https://github.com/BerriAI/litellm/pull/17979" target="_blank" rel="noopener noreferrer">PR #17979</a></li>
<li>Add OpenAI-compatible API with modify_params=True - <a href="https://github.com/BerriAI/litellm/pull/17106" target="_blank" rel="noopener noreferrer">PR #17106</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/vertex">Vertex AI</a></strong>
<ul>
<li>Add Gemini 3 Flash Preview support - <a href="https://github.com/BerriAI/litellm/pull/18164" target="_blank" rel="noopener noreferrer">PR #18164</a></li>
<li>Add reasoning support for gemini-3-flash-preview - <a href="https://github.com/BerriAI/litellm/pull/18175" target="_blank" rel="noopener noreferrer">PR #18175</a></li>
<li>Fix image edit credential source - <a href="https://github.com/BerriAI/litellm/pull/18121" target="_blank" rel="noopener noreferrer">PR #18121</a></li>
<li>Pass credentials to PredictionServiceClient for custom endpoints - <a href="https://github.com/BerriAI/litellm/pull/17757" target="_blank" rel="noopener noreferrer">PR #17757</a></li>
<li>Fix multimodal embeddings for text + base64 image combinations - <a href="https://github.com/BerriAI/litellm/pull/18172" target="_blank" rel="noopener noreferrer">PR #18172</a></li>
<li>Add OCR support for DeepSeek model - <a href="https://github.com/BerriAI/litellm/pull/17971" target="_blank" rel="noopener noreferrer">PR #17971</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/azure_ai">Azure AI</a></strong>
<ul>
<li>Add Azure Cohere 4 reranking models - <a href="https://github.com/BerriAI/litellm/pull/17961" target="_blank" rel="noopener noreferrer">PR #17961</a></li>
<li>Add Azure DeepSeek V3.2 versions - <a href="https://github.com/BerriAI/litellm/pull/18019" target="_blank" rel="noopener noreferrer">PR #18019</a></li>
<li>Return AzureAnthropicConfig for Claude models in get_provider_chat_config - <a href="https://github.com/BerriAI/litellm/pull/18086" target="_blank" rel="noopener noreferrer">PR #18086</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/fireworks_ai">Fireworks AI</a></strong>
<ul>
<li>Add reasoning param support for Fireworks AI models - <a href="https://github.com/BerriAI/litellm/pull/17967" target="_blank" rel="noopener noreferrer">PR #17967</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong>
<ul>
<li>Add Qwen 2 and Qwen 3 to get_bedrock_model_id - <a href="https://github.com/BerriAI/litellm/pull/18100" target="_blank" rel="noopener noreferrer">PR #18100</a></li>
<li>Remove ttl field when routing to bedrock - <a href="https://github.com/BerriAI/litellm/pull/18049" target="_blank" rel="noopener noreferrer">PR #18049</a></li>
<li>Add Bedrock Stability image edit models - <a href="https://github.com/BerriAI/litellm/pull/18254" target="_blank" rel="noopener noreferrer">PR #18254</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/perplexity">Perplexity</a></strong>
<ul>
<li>Use API-provided cost instead of manual calculation - <a href="https://github.com/BerriAI/litellm/pull/17887" target="_blank" rel="noopener noreferrer">PR #17887</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI</a></strong>
<ul>
<li>Add diarize model for audio transcription - <a href="https://github.com/BerriAI/litellm/pull/18117" target="_blank" rel="noopener noreferrer">PR #18117</a></li>
<li>Add gpt-image-1.5-2025-12-16 in model cost map - <a href="https://github.com/BerriAI/litellm/pull/18107" target="_blank" rel="noopener noreferrer">PR #18107</a></li>
<li>Fix cost calculation of gpt-image-1 model - <a href="https://github.com/BerriAI/litellm/pull/17966" target="_blank" rel="noopener noreferrer">PR #17966</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/github_copilot">GitHub Copilot</a></strong>
<ul>
<li>Add github_copilot model info - <a href="https://github.com/BerriAI/litellm/pull/17858" target="_blank" rel="noopener noreferrer">PR #17858</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/custom_llm_server">Custom LLM</a></strong>
<ul>
<li>Add image_edit and aimage_edit support - <a href="https://github.com/BerriAI/litellm/pull/17999" target="_blank" rel="noopener noreferrer">PR #17999</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-80-11#bug-fixes" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Gemini</a></strong>
<ul>
<li>Fix pricing for Gemini 3 Flash on Vertex AI - <a href="https://github.com/BerriAI/litellm/pull/18202" target="_blank" rel="noopener noreferrer">PR #18202</a></li>
<li>Add output_cost_per_image_token for gemini-2.5-flash-image models - <a href="https://github.com/BerriAI/litellm/pull/18156" target="_blank" rel="noopener noreferrer">PR #18156</a></li>
<li>Fix properties should be non-empty for OBJECT type - <a href="https://github.com/BerriAI/litellm/pull/18237" target="_blank" rel="noopener noreferrer">PR #18237</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/fireworks_ai">Qwen</a></strong>
<ul>
<li>Add qwen3-embedding-8b input per token price - <a href="https://github.com/BerriAI/litellm/pull/18018" target="_blank" rel="noopener noreferrer">PR #18018</a></li>
</ul>
</li>
<li><strong>General</strong>
<ul>
<li>Fix image URL handling - <a href="https://github.com/BerriAI/litellm/pull/18139" target="_blank" rel="noopener noreferrer">PR #18139</a></li>
<li>Support Signed URLs with Query Parameters in Image Processing - <a href="https://github.com/BerriAI/litellm/pull/17976" target="_blank" rel="noopener noreferrer">PR #17976</a></li>
<li>Add none to encoding_format instead of omitting it - <a href="https://github.com/BerriAI/litellm/pull/18042" target="_blank" rel="noopener noreferrer">PR #18042</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-80-11#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-80-11#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/response_api">Responses API</a></strong>
<ul>
<li>Add provider specific tools support - <a href="https://github.com/BerriAI/litellm/pull/17980" target="_blank" rel="noopener noreferrer">PR #17980</a></li>
<li>Add custom headers support - <a href="https://github.com/BerriAI/litellm/pull/18036" target="_blank" rel="noopener noreferrer">PR #18036</a></li>
<li>Fix tool calls transformation in completion bridge - <a href="https://github.com/BerriAI/litellm/pull/18226" target="_blank" rel="noopener noreferrer">PR #18226</a></li>
<li>Use list format with input_text for tool results - <a href="https://github.com/BerriAI/litellm/pull/18257" target="_blank" rel="noopener noreferrer">PR #18257</a></li>
<li>Add cost tracking in background mode - <a href="https://github.com/BerriAI/litellm/pull/18236" target="_blank" rel="noopener noreferrer">PR #18236</a></li>
<li>Fix Claude code responses API bridge errors - <a href="https://github.com/BerriAI/litellm/pull/18194" target="_blank" rel="noopener noreferrer">PR #18194</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/completion/input">Chat Completions API</a></strong>
<ul>
<li>Add support for agent skills - <a href="https://github.com/BerriAI/litellm/pull/18031" target="_blank" rel="noopener noreferrer">PR #18031</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/skills">Skills API</a></strong>
<ul>
<li>Unified Skills API works across Anthropic, Vertex, Azure, Bedrock - <a href="https://github.com/BerriAI/litellm/pull/18232" target="_blank" rel="noopener noreferrer">PR #18232</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/search/index">Search API</a></strong>
<ul>
<li>Add new RAG Search API with rerankers - <a href="https://github.com/BerriAI/litellm/pull/18217" target="_blank" rel="noopener noreferrer">PR #18217</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/interactions">Interactions API</a></strong>
<ul>
<li>Add Google Interactions API on SDK and AI Gateway - <a href="https://github.com/BerriAI/litellm/pull/18079" target="_blank" rel="noopener noreferrer">PR #18079</a>, <a href="https://github.com/BerriAI/litellm/pull/18081" target="_blank" rel="noopener noreferrer">PR #18081</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/image_edits">Image Edit API</a></strong>
<ul>
<li>Add drop_params support and fix Vertex AI config - <a href="https://github.com/BerriAI/litellm/pull/18077" target="_blank" rel="noopener noreferrer">PR #18077</a></li>
</ul>
</li>
<li><strong>General</strong>
<ul>
<li>Skip adding beta headers for Vertex AI as it is not supported - <a href="https://github.com/BerriAI/litellm/pull/18037" target="_blank" rel="noopener noreferrer">PR #18037</a></li>
<li>Fix managed files endpoint - <a href="https://github.com/BerriAI/litellm/pull/18046" target="_blank" rel="noopener noreferrer">PR #18046</a></li>
<li>Allow base_model for non-Azure providers in proxy - <a href="https://github.com/BerriAI/litellm/pull/18038" target="_blank" rel="noopener noreferrer">PR #18038</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs">Bugs<a href="https://docs.litellm.ai/release_notes/v1-80-11#bugs" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>General</strong>
<ul>
<li>Fix basemodel import in guardrail translation - <a href="https://github.com/BerriAI/litellm/pull/17977" target="_blank" rel="noopener noreferrer">PR #17977</a></li>
<li>Fix No module named 'fastapi' error - <a href="https://github.com/BerriAI/litellm/pull/18239" target="_blank" rel="noopener noreferrer">PR #18239</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-80-11#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-80-11#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li><strong>Virtual Keys</strong>
<ul>
<li>Add master key rotation for credentials table - <a href="https://github.com/BerriAI/litellm/pull/17952" target="_blank" rel="noopener noreferrer">PR #17952</a></li>
<li>Fix tag management to preserve encrypted fields in litellm_params - <a href="https://github.com/BerriAI/litellm/pull/17484" target="_blank" rel="noopener noreferrer">PR #17484</a></li>
<li>Fix key delete and regenerate permissions - <a href="https://github.com/BerriAI/litellm/pull/18214" target="_blank" rel="noopener noreferrer">PR #18214</a></li>
</ul>
</li>
<li><strong>Models + Endpoints</strong>
<ul>
<li>Add Models Conditional Rendering in UI - <a href="https://github.com/BerriAI/litellm/pull/18071" target="_blank" rel="noopener noreferrer">PR #18071</a></li>
<li>Add Health Check Model for Wildcard Model in UI - <a href="https://github.com/BerriAI/litellm/pull/18269" target="_blank" rel="noopener noreferrer">PR #18269</a></li>
<li>Auto Resolve Vector Store Embedding Model Config - <a href="https://github.com/BerriAI/litellm/pull/18167" target="_blank" rel="noopener noreferrer">PR #18167</a></li>
</ul>
</li>
<li><strong>Vector Stores</strong>
<ul>
<li>Add Milvus Vector Store UI support - <a href="https://github.com/BerriAI/litellm/pull/18030" target="_blank" rel="noopener noreferrer">PR #18030</a></li>
<li>Persist Vector Store Settings in Team Update - <a href="https://github.com/BerriAI/litellm/pull/18274" target="_blank" rel="noopener noreferrer">PR #18274</a></li>
</ul>
</li>
<li><strong>Logs &amp; Spend</strong>
<ul>
<li>Add LiteLLM Overhead to Logs - <a href="https://github.com/BerriAI/litellm/pull/18033" target="_blank" rel="noopener noreferrer">PR #18033</a></li>
<li>Show LiteLLM Overhead in Logs UI - <a href="https://github.com/BerriAI/litellm/pull/18034" target="_blank" rel="noopener noreferrer">PR #18034</a></li>
<li>Resolve Team ID to Team Alias in Usage Page - <a href="https://github.com/BerriAI/litellm/pull/18275" target="_blank" rel="noopener noreferrer">PR #18275</a></li>
<li>Fix Usage Page Top Key View Button Visibility - <a href="https://github.com/BerriAI/litellm/pull/18203" target="_blank" rel="noopener noreferrer">PR #18203</a></li>
</ul>
</li>
<li><strong>SSO &amp; Health</strong>
<ul>
<li>Add SSO Readiness Health Check - <a href="https://github.com/BerriAI/litellm/pull/18078" target="_blank" rel="noopener noreferrer">PR #18078</a></li>
<li>Fix /health/test_connection to resolve env variables like /chat/completions - <a href="https://github.com/BerriAI/litellm/pull/17752" target="_blank" rel="noopener noreferrer">PR #17752</a></li>
</ul>
</li>
<li><strong>CloudZero</strong>
<ul>
<li>Add CloudZero Cost Tracking UI - <a href="https://github.com/BerriAI/litellm/pull/18163" target="_blank" rel="noopener noreferrer">PR #18163</a></li>
<li>Add Delete CloudZero Settings Route and UI - <a href="https://github.com/BerriAI/litellm/pull/18168" target="_blank" rel="noopener noreferrer">PR #18168</a>, <a href="https://github.com/BerriAI/litellm/pull/18170" target="_blank" rel="noopener noreferrer">PR #18170</a></li>
</ul>
</li>
<li><strong>General</strong>
<ul>
<li>Update UI path handling for non-root Docker - <a href="https://github.com/BerriAI/litellm/pull/17989" target="_blank" rel="noopener noreferrer">PR #17989</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs-1">Bugs<a href="https://docs.litellm.ai/release_notes/v1-80-11#bugs-1" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>UI Fixes</strong>
<ul>
<li>Fix Login Page Failed To Parse JSON Error - <a href="https://github.com/BerriAI/litellm/pull/18159" target="_blank" rel="noopener noreferrer">PR #18159</a></li>
<li>Fix new user route user_id collision handling - <a href="https://github.com/BerriAI/litellm/pull/17559" target="_blank" rel="noopener noreferrer">PR #17559</a></li>
<li>Fix Callback Environment Variables Casing - <a href="https://github.com/BerriAI/litellm/pull/17912" target="_blank" rel="noopener noreferrer">PR #17912</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="ai-integrations">AI Integrations<a href="https://docs.litellm.ai/release_notes/v1-80-11#ai-integrations" class="hash-link" aria-label="Direct link to AI Integrations" title="Direct link to AI Integrations">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="logging">Logging<a href="https://docs.litellm.ai/release_notes/v1-80-11#logging" class="hash-link" aria-label="Direct link to Logging" title="Direct link to Logging">​</a></h3>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/observability/azure_sentinel">Azure Sentinel</a></strong>
<ul>
<li>Add new Azure Sentinel Logger integration - <a href="https://github.com/BerriAI/litellm/pull/18146" target="_blank" rel="noopener noreferrer">PR #18146</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging#prometheus">Prometheus</a></strong>
<ul>
<li>Add extraction of top level metadata for custom labels - <a href="https://github.com/BerriAI/litellm/pull/18087" target="_blank" rel="noopener noreferrer">PR #18087</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging#langfuse">Langfuse</a></strong>
<ul>
<li>Fix not working log_failure_event - <a href="https://github.com/BerriAI/litellm/pull/18234" target="_blank" rel="noopener noreferrer">PR #18234</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/observability/phoenix_integration">Arize Phoenix</a></strong>
<ul>
<li>Fix nested spans - <a href="https://github.com/BerriAI/litellm/pull/18102" target="_blank" rel="noopener noreferrer">PR #18102</a></li>
</ul>
</li>
<li><strong>General</strong>
<ul>
<li>Change extra_headers to additional_headers - <a href="https://github.com/BerriAI/litellm/pull/17950" target="_blank" rel="noopener noreferrer">PR #17950</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails">Guardrails<a href="https://docs.litellm.ai/release_notes/v1-80-11#guardrails" class="hash-link" aria-label="Direct link to Guardrails" title="Direct link to Guardrails">​</a></h3>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/litellm_content_filter">LiteLLM Content Filter</a></strong>
<ul>
<li>Add built-in guardrails for harmful content, bias, etc. - <a href="https://github.com/BerriAI/litellm/pull/18029" target="_blank" rel="noopener noreferrer">PR #18029</a></li>
<li>Add support for running content filters on images - <a href="https://github.com/BerriAI/litellm/pull/18044" target="_blank" rel="noopener noreferrer">PR #18044</a></li>
<li>Add support for Brazil PII field - <a href="https://github.com/BerriAI/litellm/pull/18076" target="_blank" rel="noopener noreferrer">PR #18076</a></li>
<li>Add configurable guardrail options for content filtering - <a href="https://github.com/BerriAI/litellm/pull/18007" target="_blank" rel="noopener noreferrer">PR #18007</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/adding_provider/generic_guardrail_api">Guardrails API</a></strong>
<ul>
<li>Support LLM tool call response checks on <code>/chat/completions</code>, <code>/v1/responses</code>, <code>/v1/messages</code> - <a href="https://github.com/BerriAI/litellm/pull/17619" target="_blank" rel="noopener noreferrer">PR #17619</a></li>
<li>Add guardrails load balancing - <a href="https://github.com/BerriAI/litellm/pull/18181" target="_blank" rel="noopener noreferrer">PR #18181</a></li>
<li>Fix guardrails for passthrough endpoint - <a href="https://github.com/BerriAI/litellm/pull/18109" target="_blank" rel="noopener noreferrer">PR #18109</a></li>
<li>Add headers to metadata for guardrails on pass-through endpoints - <a href="https://github.com/BerriAI/litellm/pull/17992" target="_blank" rel="noopener noreferrer">PR #17992</a></li>
<li>Various fixes for guardrail on OpenRouter models - <a href="https://github.com/BerriAI/litellm/pull/18085" target="_blank" rel="noopener noreferrer">PR #18085</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/lakera_ai">Lakera</a></strong>
<ul>
<li>Add monitor mode for Lakera - <a href="https://github.com/BerriAI/litellm/pull/18084" target="_blank" rel="noopener noreferrer">PR #18084</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/pillar_security">Pillar Security</a></strong>
<ul>
<li>Add masking support and MCP call support - <a href="https://github.com/BerriAI/litellm/pull/17959" target="_blank" rel="noopener noreferrer">PR #17959</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/bedrock">Bedrock Guardrails</a></strong>
<ul>
<li>Add support for Bedrock image guardrails - <a href="https://github.com/BerriAI/litellm/pull/18115" target="_blank" rel="noopener noreferrer">PR #18115</a></li>
<li>Guardrails block action takes precedence over masking - <a href="https://github.com/BerriAI/litellm/pull/17968" target="_blank" rel="noopener noreferrer">PR #17968</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="secret-managers">Secret Managers<a href="https://docs.litellm.ai/release_notes/v1-80-11#secret-managers" class="hash-link" aria-label="Direct link to Secret Managers" title="Direct link to Secret Managers">​</a></h3>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/secret_managers/hashicorp_vault">HashiCorp Vault</a></strong>
<ul>
<li>Add documentation for configurable Vault mount - <a href="https://github.com/BerriAI/litellm/pull/18082" target="_blank" rel="noopener noreferrer">PR #18082</a></li>
<li>Add per-team Vault configuration - <a href="https://github.com/BerriAI/litellm/pull/18150" target="_blank" rel="noopener noreferrer">PR #18150</a></li>
</ul>
</li>
<li><strong>UI</strong>
<ul>
<li>Add secret manager settings controls to team management UI - <a href="https://github.com/BerriAI/litellm/pull/18149" target="_blank" rel="noopener noreferrer">PR #18149</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-80-11#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li><strong>Email Budget Alerts</strong> - Send email notifications when budgets are reached - <a href="https://github.com/BerriAI/litellm/pull/17995" target="_blank" rel="noopener noreferrer">PR #17995</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-gateway">MCP Gateway<a href="https://docs.litellm.ai/release_notes/v1-80-11#mcp-gateway" class="hash-link" aria-label="Direct link to MCP Gateway" title="Direct link to MCP Gateway">​</a></h2>
<ul>
<li><strong>Auth Header Propagation</strong> - Add MCP auth header propagation - <a href="https://github.com/BerriAI/litellm/pull/17963" target="_blank" rel="noopener noreferrer">PR #17963</a></li>
<li><strong>Fix deepcopy error</strong> - Fix MCP tool call deepcopy error when processing requests - <a href="https://github.com/BerriAI/litellm/pull/18010" target="_blank" rel="noopener noreferrer">PR #18010</a></li>
<li><strong>Fix list tool</strong> - Fix MCP list_tools not working without database connection - <a href="https://github.com/BerriAI/litellm/pull/18161" target="_blank" rel="noopener noreferrer">PR #18161</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="agent-gateway-a2a">Agent Gateway (A2A)<a href="https://docs.litellm.ai/release_notes/v1-80-11#agent-gateway-a2a" class="hash-link" aria-label="Direct link to Agent Gateway (A2A)" title="Direct link to Agent Gateway (A2A)">​</a></h2>
<ul>
<li><strong>New Provider: Agent Gateway</strong> - Add pydantic ai agents support - <a href="https://github.com/BerriAI/litellm/pull/18013" target="_blank" rel="noopener noreferrer">PR #18013</a></li>
<li><strong>VertexAI Agent Engine</strong> - Add Vertex AI Agent Engine provider - <a href="https://github.com/BerriAI/litellm/pull/18014" target="_blank" rel="noopener noreferrer">PR #18014</a></li>
<li><strong>Fix model extraction</strong> - Fix get_model_from_request() to extract model ID from Vertex AI passthrough URLs - <a href="https://github.com/BerriAI/litellm/pull/18097" target="_blank" rel="noopener noreferrer">PR #18097</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements">Performance / Loadbalancing / Reliability improvements<a href="https://docs.litellm.ai/release_notes/v1-80-11#performance--loadbalancing--reliability-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements" title="Direct link to Performance / Loadbalancing / Reliability improvements">​</a></h2>
<ul>
<li><strong>Lazy Imports</strong> - Use per-attribute lazy imports and extract shared constants - <a href="https://github.com/BerriAI/litellm/pull/17994" target="_blank" rel="noopener noreferrer">PR #17994</a></li>
<li><strong>Lazy Load HTTP Handlers</strong> - Lazy load http handlers - <a href="https://github.com/BerriAI/litellm/pull/17997" target="_blank" rel="noopener noreferrer">PR #17997</a></li>
<li><strong>Lazy Load Caches</strong> - Lazy load caches - <a href="https://github.com/BerriAI/litellm/pull/18001" target="_blank" rel="noopener noreferrer">PR #18001</a></li>
<li><strong>Lazy Load Types</strong> - Lazy load bedrock types, .types.utils, GuardrailItem - <a href="https://github.com/BerriAI/litellm/pull/18053" target="_blank" rel="noopener noreferrer">PR #18053</a>, <a href="https://github.com/BerriAI/litellm/pull/18054" target="_blank" rel="noopener noreferrer">PR #18054</a>, <a href="https://github.com/BerriAI/litellm/pull/18072" target="_blank" rel="noopener noreferrer">PR #18072</a></li>
<li><strong>Lazy Load Configs</strong> - Lazy load 41 configuration classes - <a href="https://github.com/BerriAI/litellm/pull/18267" target="_blank" rel="noopener noreferrer">PR #18267</a></li>
<li><strong>Lazy Load Client Decorators</strong> - Lazy load heavy client decorator imports - <a href="https://github.com/BerriAI/litellm/pull/18064" target="_blank" rel="noopener noreferrer">PR #18064</a></li>
<li><strong>Prisma Build Time</strong> - Download Prisma binaries at build time instead of runtime for security restricted environments - <a href="https://github.com/BerriAI/litellm/pull/17695" target="_blank" rel="noopener noreferrer">PR #17695</a></li>
<li><strong>Docker Alpine</strong> - Add libsndfile to Alpine image for ARM64 audio processing - <a href="https://github.com/BerriAI/litellm/pull/18092" target="_blank" rel="noopener noreferrer">PR #18092</a></li>
<li><strong>Security</strong> - Prevent LiteLLM API key leakage on /health endpoint failures - <a href="https://github.com/BerriAI/litellm/pull/18133" target="_blank" rel="noopener noreferrer">PR #18133</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates">Documentation Updates<a href="https://docs.litellm.ai/release_notes/v1-80-11#documentation-updates" class="hash-link" aria-label="Direct link to Documentation Updates" title="Direct link to Documentation Updates">​</a></h2>
<ul>
<li><strong>SAP Docs</strong> - Update SAP documentation - <a href="https://github.com/BerriAI/litellm/pull/17974" target="_blank" rel="noopener noreferrer">PR #17974</a></li>
<li><strong>Pydantic AI Agents</strong> - Add docs on using pydantic ai agents with LiteLLM A2A gateway - <a href="https://github.com/BerriAI/litellm/pull/18026" target="_blank" rel="noopener noreferrer">PR #18026</a></li>
<li><strong>Vertex AI Agent Engine</strong> - Add Vertex AI Agent Engine documentation - <a href="https://github.com/BerriAI/litellm/pull/18027" target="_blank" rel="noopener noreferrer">PR #18027</a></li>
<li><strong>Router Order</strong> - Add router order parameter documentation - <a href="https://github.com/BerriAI/litellm/pull/18045" target="_blank" rel="noopener noreferrer">PR #18045</a></li>
<li><strong>Secret Manager Settings</strong> - Improve secret manager settings documentation - <a href="https://github.com/BerriAI/litellm/pull/18235" target="_blank" rel="noopener noreferrer">PR #18235</a></li>
<li><strong>Gemini 3 Flash</strong> - Add version requirement in Gemini 3 Flash blog - <a href="https://github.com/BerriAI/litellm/pull/18227" target="_blank" rel="noopener noreferrer">PR #18227</a></li>
<li><strong>README</strong> - Expand Responses API section and update endpoints - <a href="https://github.com/BerriAI/litellm/pull/17354" target="_blank" rel="noopener noreferrer">PR #17354</a></li>
<li><strong>Amazon Nova</strong> - Add Amazon Nova to sidebar and supported models - <a href="https://github.com/BerriAI/litellm/pull/18220" target="_blank" rel="noopener noreferrer">PR #18220</a></li>
<li><strong>Benchmarks</strong> - Add infrastructure recommendations to benchmarks documentation - <a href="https://github.com/BerriAI/litellm/pull/18264" target="_blank" rel="noopener noreferrer">PR #18264</a></li>
<li><strong>Broken Links</strong> - Fix broken link corrections - <a href="https://github.com/BerriAI/litellm/pull/18104" target="_blank" rel="noopener noreferrer">PR #18104</a></li>
<li><strong>README Fixes</strong> - Various README improvements - <a href="https://github.com/BerriAI/litellm/pull/18206" target="_blank" rel="noopener noreferrer">PR #18206</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="infrastructure--cicd">Infrastructure / CI/CD<a href="https://docs.litellm.ai/release_notes/v1-80-11#infrastructure--cicd" class="hash-link" aria-label="Direct link to Infrastructure / CI/CD" title="Direct link to Infrastructure / CI/CD">​</a></h2>
<ul>
<li><strong>PR Templates</strong> - Add LiteLLM team PR template and CI/CD rules - <a href="https://github.com/BerriAI/litellm/pull/17983" target="_blank" rel="noopener noreferrer">PR #17983</a>, <a href="https://github.com/BerriAI/litellm/pull/17985" target="_blank" rel="noopener noreferrer">PR #17985</a></li>
<li><strong>Issue Labeling</strong> - Improve issue labeling with component dropdown and more provider keywords - <a href="https://github.com/BerriAI/litellm/pull/17957" target="_blank" rel="noopener noreferrer">PR #17957</a></li>
<li><strong>PR Template Cleanup</strong> - Remove redundant fields from PR template - <a href="https://github.com/BerriAI/litellm/pull/17956" target="_blank" rel="noopener noreferrer">PR #17956</a></li>
<li><strong>Dependencies</strong> - Bump altcha-lib from 1.3.0 to 1.4.1 - <a href="https://github.com/BerriAI/litellm/pull/18017" target="_blank" rel="noopener noreferrer">PR #18017</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-80-11#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@dongbin-lunark made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17757" target="_blank" rel="noopener noreferrer">PR #17757</a></li>
<li>@qdrddr made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18004" target="_blank" rel="noopener noreferrer">PR #18004</a></li>
<li>@donicrosby made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17962" target="_blank" rel="noopener noreferrer">PR #17962</a></li>
<li>@NicolaivdSmagt made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17992" target="_blank" rel="noopener noreferrer">PR #17992</a></li>
<li>@Reapor-Yurnero made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18085" target="_blank" rel="noopener noreferrer">PR #18085</a></li>
<li>@jk-f5 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18086" target="_blank" rel="noopener noreferrer">PR #18086</a></li>
<li>@castrapel made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18077" target="_blank" rel="noopener noreferrer">PR #18077</a></li>
<li>@dtikhonov made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17484" target="_blank" rel="noopener noreferrer">PR #17484</a></li>
<li>@opleonnn made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18175" target="_blank" rel="noopener noreferrer">PR #18175</a></li>
<li>@eurogig made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/18084" target="_blank" rel="noopener noreferrer">PR #18084</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog">Full Changelog<a href="https://docs.litellm.ai/release_notes/v1-80-11#full-changelog" class="hash-link" aria-label="Direct link to Full Changelog" title="Direct link to Full Changelog">​</a></h2>
<p><strong><a href="https://github.com/BerriAI/litellm/compare/v1.80.10-nightly...v1.80.11" target="_blank" rel="noopener noreferrer">View complete changelog on GitHub</a></strong></p>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[[Preview] v1.80.10.rc.1 - Agent Gateway: Azure Foundry & Bedrock AgentCore]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-80-10</id>
        <link href="https://docs.litellm.ai/release_notes/v1-80-10"/>
        <updated>2025-12-13T10:00:00.000Z</updated>
        <summary type="html"><![CDATA[Deploy this version]]></summary>
        <content type="html"><![CDATA[<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-80-10#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">docker.litellm.ai/berriai/litellm:v1.80.10.rc.1</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.80.10</span><br></span></code></pre></div></div></div></div></div>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-80-10#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<ul>
<li><strong>Agent (A2A) Gateway with Cost Tracking</strong> - <a href="https://docs.litellm.ai/docs/a2a_cost_tracking">Track agent costs per query, per token pricing, and view agent usage in the dashboard</a></li>
<li><strong>2 New Agent Providers</strong> - <a href="https://docs.litellm.ai/docs/providers/langgraph">LangGraph Agents</a> and <a href="https://docs.litellm.ai/docs/providers/azure_ai_agents">Azure AI Foundry Agents</a> for agentic workflows</li>
<li><strong>New Provider: SAP Gen AI Hub</strong> - <a href="https://docs.litellm.ai/docs/providers/sap">Full support for SAP Generative AI Hub with chat completions</a></li>
<li><strong>New Bedrock Writer Models</strong> - Add Palmyra-X4 and Palmyra-X5 models on Bedrock</li>
<li><strong>OpenAI GPT-5.2 Models</strong> - Full support for GPT-5.2, GPT-5.2-pro, and Azure GPT-5.2 models with reasoning support</li>
<li><strong>227 New Fireworks AI Models</strong> - Comprehensive model coverage for Fireworks AI platform</li>
<li><strong>MCP Support on /chat/completions</strong> - <a href="https://docs.litellm.ai/docs/mcp">Use MCP servers directly via chat completions endpoint</a></li>
<li><strong>Performance Improvements</strong> - Reduced memory leaks by 50%</li>
</ul>
<hr>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="agent-gateway---4-new-agent-providers">Agent Gateway - 4 New Agent Providers<a href="https://docs.litellm.ai/release_notes/v1-80-10#agent-gateway---4-new-agent-providers" class="hash-link" aria-label="Direct link to Agent Gateway - 4 New Agent Providers" title="Direct link to Agent Gateway - 4 New Agent Providers">​</a></h3>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAFCAYAAAB8ZH1oAAAACXBIWXMAACxLAAAsSwGlPZapAAAAtklEQVR4nAXB207CMACA4d4qiLJDt8ZWYK47wDo5uGhDwMVENDEeQoje+f5v8ft9ItYtkW45DyxZ5amXPU33RtV9EpYfyOaHxJ0Qg8ByFuTEN0sKt2dae8z8icn6G/PwR+J+kc0JMVItw/QOlXtmrie1nsRuUYsDenMkqt4J7StiPN1xNeu5rp/JVi8k5R5Z7IjtFpl7Ls0jI90hInsgrL9QRU9W3mNuV6iJQ+oF0jgu0oahnPMPeMlJkr5fegQAAAAASUVORK5CYII=&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="331"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/a2a_gateway2.f1b1005.640.png" srcset="/assets/ideal-img/a2a_gateway2.f1b1005.640.png 640w,/assets/ideal-img/a2a_gateway2.c658f49.1920.png 1920w" width="640" height="331"></noscript></div>
<br>
<p>This release adds support for agents from the following providers:</p>
<ul>
<li><strong>LangGraph Agents</strong> - Deploy and manage LangGraph-based agents</li>
<li><strong>Azure AI Foundry Agents</strong> - Enterprise agent deployments on Azure</li>
<li><strong>Bedrock AgentCore</strong> - AWS Bedrock agent integration</li>
<li><strong>A2A Agents</strong> - Agent-to-Agent protocol support</li>
</ul>
<p>AI Gateway admins can now add agents from any of these providers, and developers can invoke them through a unified interface using the A2A protocol.</p>
<p>For all agent requests running through the AI Gateway, LiteLLM automatically tracks request/response logs, cost, and token usage.</p>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="agent-a2a-usage-ui">Agent (A2A) Usage UI<a href="https://docs.litellm.ai/release_notes/v1-80-10#agent-a2a-usage-ui" class="hash-link" aria-label="Direct link to Agent (A2A) Usage UI" title="Direct link to Agent (A2A) Usage UI">​</a></h3>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAFCAYAAAB8ZH1oAAAACXBIWXMAABYlAAAWJQFJUiTwAAAAhklEQVR4nE2NWwrCQBRDu3/wx624AJeh9m+qqPgxKrWt9zH3yLQKBg4kEJIm50yl73tE5IuiulCzqtCklGjblq47MgwjUgseEEEpgbtzfxiNmfFjHCfCHXeb16pXNUSc5v+CKKTXm/1jAAKxQpRlfS7Wu3DjNinrw4XV7szmlNlen1RNVvgAjSnA6BmdHE4AAAAASUVORK5CYII=&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="334"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/agent_usage.7ed2cbf.640.png" srcset="/assets/ideal-img/agent_usage.7ed2cbf.640.png 640w,/assets/ideal-img/agent_usage.a83b108.1920.png 1920w" width="640" height="334"></noscript></div>
<p>Users can now filter usage statistics by agents, providing the same granular filtering capabilities available for teams, organizations, and customers.</p>
<p><strong>Details:</strong></p>
<ul>
<li>Filter usage analytics, spend logs, and activity metrics by agent ID</li>
<li>View breakdowns on a per-agent basis</li>
<li>Consistent filtering experience across all usage and analytics views</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-providers-and-endpoints">New Providers and Endpoints<a href="https://docs.litellm.ai/release_notes/v1-80-10#new-providers-and-endpoints" class="hash-link" aria-label="Direct link to New Providers and Endpoints" title="Direct link to New Providers and Endpoints">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="new-providers-5-new-providers">New Providers (5 new providers)<a href="https://docs.litellm.ai/release_notes/v1-80-10#new-providers-5-new-providers" class="hash-link" aria-label="Direct link to New Providers (5 new providers)" title="Direct link to New Providers (5 new providers)">​</a></h3>
<table><thead><tr><th>Provider</th><th>Supported LiteLLM Endpoints</th><th>Description</th></tr></thead><tbody><tr><td><a href="https://docs.litellm.ai/docs/providers/sap">SAP Gen AI Hub</a></td><td><code>/chat/completions</code>, <code>/messages</code>, <code>/responses</code></td><td>SAP Generative AI Hub integration for enterprise AI</td></tr><tr><td><a href="https://docs.litellm.ai/docs/providers/langgraph">LangGraph</a></td><td><code>/chat/completions</code>, <code>/messages</code>, <code>/responses</code>, <code>/a2a</code></td><td>LangGraph agents for agentic workflows</td></tr><tr><td><a href="https://docs.litellm.ai/docs/providers/azure_ai_agents">Azure AI Foundry Agents</a></td><td><code>/chat/completions</code>, <code>/messages</code>, <code>/responses</code>, <code>/a2a</code></td><td>Azure AI Foundry Agents for enterprise agent deployments</td></tr><tr><td><a href="https://docs.litellm.ai/docs/providers/voyage">Voyage AI Rerank</a></td><td><code>/rerank</code></td><td>Voyage AI rerank models support</td></tr><tr><td><a href="https://docs.litellm.ai/docs/providers/fireworks_ai">Fireworks AI Rerank</a></td><td><code>/rerank</code></td><td>Fireworks AI rerank endpoint support</td></tr></tbody></table>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="new-llm-api-endpoints-4-new-endpoints">New LLM API Endpoints (4 new endpoints)<a href="https://docs.litellm.ai/release_notes/v1-80-10#new-llm-api-endpoints-4-new-endpoints" class="hash-link" aria-label="Direct link to New LLM API Endpoints (4 new endpoints)" title="Direct link to New LLM API Endpoints (4 new endpoints)">​</a></h3>
<table><thead><tr><th>Endpoint</th><th>Method</th><th>Description</th><th>Documentation</th></tr></thead><tbody><tr><td><code>/containers/{id}/files</code></td><td>GET</td><td>List files in a container</td><td><a href="https://docs.litellm.ai/docs/container_files">Docs</a></td></tr><tr><td><code>/containers/{id}/files/{file_id}</code></td><td>GET</td><td>Retrieve container file metadata</td><td><a href="https://docs.litellm.ai/docs/container_files">Docs</a></td></tr><tr><td><code>/containers/{id}/files/{file_id}</code></td><td>DELETE</td><td>Delete a file from a container</td><td><a href="https://docs.litellm.ai/docs/container_files">Docs</a></td></tr><tr><td><code>/containers/{id}/files/{file_id}/content</code></td><td>GET</td><td>Retrieve container file content</td><td><a href="https://docs.litellm.ai/docs/container_files">Docs</a></td></tr></tbody></table>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-80-10#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support-270-new-models">New Model Support (270+ new models)<a href="https://docs.litellm.ai/release_notes/v1-80-10#new-model-support-270-new-models" class="hash-link" aria-label="Direct link to New Model Support (270+ new models)" title="Direct link to New Model Support (270+ new models)">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th><th>Features</th></tr></thead><tbody><tr><td>OpenAI</td><td><code>gpt-5.2</code></td><td>400K</td><td>$1.75</td><td>$14.00</td><td>Reasoning, vision, PDF, caching</td></tr><tr><td>OpenAI</td><td><code>gpt-5.2-pro</code></td><td>400K</td><td>$21.00</td><td>$168.00</td><td>Reasoning, web search, vision</td></tr><tr><td>Azure</td><td><code>azure/gpt-5.2</code></td><td>400K</td><td>$1.75</td><td>$14.00</td><td>Reasoning, vision, PDF, caching</td></tr><tr><td>Azure</td><td><code>azure/gpt-5.2-pro</code></td><td>400K</td><td>$21.00</td><td>$168.00</td><td>Reasoning, web search</td></tr><tr><td>Bedrock</td><td><code>us.writer.palmyra-x4-v1:0</code></td><td>128K</td><td>$2.50</td><td>$10.00</td><td>Function calling, PDF input</td></tr><tr><td>Bedrock</td><td><code>us.writer.palmyra-x5-v1:0</code></td><td>1M</td><td>$0.60</td><td>$6.00</td><td>Function calling, PDF input</td></tr><tr><td>Bedrock</td><td><code>eu.anthropic.claude-opus-4-5-20251101-v1:0</code></td><td>200K</td><td>$5.00</td><td>$25.00</td><td>Reasoning, computer use, vision</td></tr><tr><td>Bedrock</td><td><code>google.gemma-3-12b-it</code></td><td>128K</td><td>$0.10</td><td>$0.30</td><td>Audio input</td></tr><tr><td>Bedrock</td><td><code>moonshot.kimi-k2-thinking</code></td><td>128K</td><td>$0.60</td><td>$2.50</td><td>Reasoning</td></tr><tr><td>Bedrock</td><td><code>nvidia.nemotron-nano-12b-v2</code></td><td>128K</td><td>$0.20</td><td>$0.60</td><td>Vision</td></tr><tr><td>Bedrock</td><td><code>qwen.qwen3-next-80b-a3b</code></td><td>128K</td><td>$0.15</td><td>$1.20</td><td>Function calling</td></tr><tr><td>Vertex AI</td><td><code>vertex_ai/deepseek-ai/deepseek-v3.2-maas</code></td><td>164K</td><td>$0.56</td><td>$1.68</td><td>Reasoning, caching</td></tr><tr><td>Mistral</td><td><code>mistral/codestral-2508</code></td><td>256K</td><td>$0.30</td><td>$0.90</td><td>Function calling</td></tr><tr><td>Mistral</td><td><code>mistral/devstral-2512</code></td><td>256K</td><td>$0.40</td><td>$2.00</td><td>Function calling</td></tr><tr><td>Mistral</td><td><code>mistral/labs-devstral-small-2512</code></td><td>256K</td><td>$0.10</td><td>$0.30</td><td>Function calling</td></tr><tr><td>Cerebras</td><td><code>cerebras/zai-glm-4.6</code></td><td>128K</td><td>-</td><td>-</td><td>Chat completions</td></tr><tr><td>NVIDIA NIM</td><td><code>nvidia_nim/ranking/nvidia/llama-3.2-nv-rerankqa-1b-v2</code></td><td>-</td><td>Free</td><td>Free</td><td>Rerank</td></tr><tr><td>Voyage</td><td><code>voyage/rerank-2.5</code></td><td>32K</td><td>$0.05/1K tokens</td><td>-</td><td>Rerank</td></tr><tr><td>Fireworks AI</td><td>227 new models</td><td>Various</td><td>Various</td><td>Various</td><td>Full model catalog</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-80-10#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI</a></strong>
<ul>
<li>Add support for OpenAI GPT-5.2 models with reasoning_effort='xhigh' - <a href="https://github.com/BerriAI/litellm/pull/17836" target="_blank" rel="noopener noreferrer">PR #17836</a>, <a href="https://github.com/BerriAI/litellm/pull/17875" target="_blank" rel="noopener noreferrer">PR #17875</a></li>
<li>Include 'user' param for responses API models - <a href="https://github.com/BerriAI/litellm/pull/17648" target="_blank" rel="noopener noreferrer">PR #17648</a></li>
<li>Use optimized async http client for text completions - <a href="https://github.com/BerriAI/litellm/pull/17831" target="_blank" rel="noopener noreferrer">PR #17831</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/azure">Azure</a></strong>
<ul>
<li>Add Azure GPT-5.2 models support - <a href="https://github.com/BerriAI/litellm/pull/17866" target="_blank" rel="noopener noreferrer">PR #17866</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/azure_ai">Azure AI</a></strong>
<ul>
<li>Fix Azure AI Anthropic api-key header and passthrough cost calculation - <a href="https://github.com/BerriAI/litellm/pull/17656" target="_blank" rel="noopener noreferrer">PR #17656</a></li>
<li>Remove unsupported params from Azure AI Anthropic requests - <a href="https://github.com/BerriAI/litellm/pull/17822" target="_blank" rel="noopener noreferrer">PR #17822</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong>
<ul>
<li>Prevent duplicate tool_result blocks with same tool - <a href="https://github.com/BerriAI/litellm/pull/17632" target="_blank" rel="noopener noreferrer">PR #17632</a></li>
<li>Handle partial JSON chunks in streaming responses - <a href="https://github.com/BerriAI/litellm/pull/17493" target="_blank" rel="noopener noreferrer">PR #17493</a></li>
<li>Preserve server_tool_use and web_search_tool_result in multi-turn conversations - <a href="https://github.com/BerriAI/litellm/pull/17746" target="_blank" rel="noopener noreferrer">PR #17746</a></li>
<li>Capture web_search_tool_result in streaming for multi-turn conversations - <a href="https://github.com/BerriAI/litellm/pull/17798" target="_blank" rel="noopener noreferrer">PR #17798</a></li>
<li>Add retrieve batches and retrieve file content support - <a href="https://github.com/BerriAI/litellm/pull/17700" target="_blank" rel="noopener noreferrer">PR #17700</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong>
<ul>
<li>Add new Bedrock OSS models to model list - <a href="https://github.com/BerriAI/litellm/pull/17638" target="_blank" rel="noopener noreferrer">PR #17638</a></li>
<li>Add Bedrock Writer models (Palmyra-X4, Palmyra-X5) - <a href="https://github.com/BerriAI/litellm/pull/17685" target="_blank" rel="noopener noreferrer">PR #17685</a></li>
<li>Add EU Claude Opus 4.5 model - <a href="https://github.com/BerriAI/litellm/pull/17897" target="_blank" rel="noopener noreferrer">PR #17897</a></li>
<li>Add serviceTier support for Converse API - <a href="https://github.com/BerriAI/litellm/pull/17810" target="_blank" rel="noopener noreferrer">PR #17810</a></li>
<li>Fix header forwarding with custom API for Bedrock embeddings - <a href="https://github.com/BerriAI/litellm/pull/17872" target="_blank" rel="noopener noreferrer">PR #17872</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Gemini</a></strong>
<ul>
<li>Add support for computer use for Gemini - <a href="https://github.com/BerriAI/litellm/pull/17756" target="_blank" rel="noopener noreferrer">PR #17756</a></li>
<li>Handle context window errors - <a href="https://github.com/BerriAI/litellm/pull/17751" target="_blank" rel="noopener noreferrer">PR #17751</a></li>
<li>Add speechConfig to GenerationConfig for Gemini TTS - <a href="https://github.com/BerriAI/litellm/pull/17851" target="_blank" rel="noopener noreferrer">PR #17851</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/vertex">Vertex AI</a></strong>
<ul>
<li>Add DeepSeek-V3.2 model support - <a href="https://github.com/BerriAI/litellm/pull/17770" target="_blank" rel="noopener noreferrer">PR #17770</a></li>
<li>Preserve systemInstructions for generate content request - <a href="https://github.com/BerriAI/litellm/pull/17803" target="_blank" rel="noopener noreferrer">PR #17803</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/mistral">Mistral</a></strong>
<ul>
<li>Add Codestral 2508, Devstral 2512 models - <a href="https://github.com/BerriAI/litellm/pull/17801" target="_blank" rel="noopener noreferrer">PR #17801</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/cerebras">Cerebras</a></strong>
<ul>
<li>Add zai-glm-4.6 model support - <a href="https://github.com/BerriAI/litellm/pull/17683" target="_blank" rel="noopener noreferrer">PR #17683</a></li>
<li>Fix context window errors not recognized - <a href="https://github.com/BerriAI/litellm/pull/17587" target="_blank" rel="noopener noreferrer">PR #17587</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/deepseek">DeepSeek</a></strong>
<ul>
<li>Add native support for thinking and reasoning_effort params - <a href="https://github.com/BerriAI/litellm/pull/17712" target="_blank" rel="noopener noreferrer">PR #17712</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/nvidia_nim_rerank">NVIDIA NIM Rerank</a></strong>
<ul>
<li>Add llama-3.2-nv-rerankqa-1b-v2 rerank model - <a href="https://github.com/BerriAI/litellm/pull/17670" target="_blank" rel="noopener noreferrer">PR #17670</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/fireworks_ai">Fireworks AI</a></strong>
<ul>
<li>Add 227 new Fireworks AI models - <a href="https://github.com/BerriAI/litellm/pull/17692" target="_blank" rel="noopener noreferrer">PR #17692</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/dashscope">Dashscope</a></strong>
<ul>
<li>Fix default base_url error - <a href="https://github.com/BerriAI/litellm/pull/17584" target="_blank" rel="noopener noreferrer">PR #17584</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-80-10#bug-fixes" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong>
<ul>
<li>Fix missing content in Anthropic to OpenAI conversion - <a href="https://github.com/BerriAI/litellm/pull/17693" target="_blank" rel="noopener noreferrer">PR #17693</a></li>
<li>Avoid error when we have just the tool_calls in input - <a href="https://github.com/BerriAI/litellm/pull/17753" target="_blank" rel="noopener noreferrer">PR #17753</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/azure">Azure</a></strong>
<ul>
<li>Fix error about encoding video id for Azure - <a href="https://github.com/BerriAI/litellm/pull/17708" target="_blank" rel="noopener noreferrer">PR #17708</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/azure_ai">Azure AI</a></strong>
<ul>
<li>Fix LLM provider for azure_ai in model map - <a href="https://github.com/BerriAI/litellm/pull/17805" target="_blank" rel="noopener noreferrer">PR #17805</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/watsonx">Watsonx</a></strong>
<ul>
<li>Fix Watsonx Audio Transcription to only send supported params to API - <a href="https://github.com/BerriAI/litellm/pull/17840" target="_blank" rel="noopener noreferrer">PR #17840</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/routing">Router</a></strong>
<ul>
<li>Handle tools=None in completion requests - <a href="https://github.com/BerriAI/litellm/pull/17684" target="_blank" rel="noopener noreferrer">PR #17684</a></li>
<li>Add minimum request threshold for error rate cooldown - <a href="https://github.com/BerriAI/litellm/pull/17464" target="_blank" rel="noopener noreferrer">PR #17464</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-80-10#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-80-10#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/response_api">Responses API</a></strong>
<ul>
<li>Add usage details in responses usage object - <a href="https://github.com/BerriAI/litellm/pull/17641" target="_blank" rel="noopener noreferrer">PR #17641</a></li>
<li>Fix error for response API polling - <a href="https://github.com/BerriAI/litellm/pull/17654" target="_blank" rel="noopener noreferrer">PR #17654</a></li>
<li>Fix streaming tool_calls being dropped when text + tool_calls - <a href="https://github.com/BerriAI/litellm/pull/17652" target="_blank" rel="noopener noreferrer">PR #17652</a></li>
<li>Transform image content in tool results for Responses API - <a href="https://github.com/BerriAI/litellm/pull/17799" target="_blank" rel="noopener noreferrer">PR #17799</a></li>
<li>Fix responses api not applying tpm rate limits on api keys - <a href="https://github.com/BerriAI/litellm/pull/17707" target="_blank" rel="noopener noreferrer">PR #17707</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/containers">Containers API</a></strong>
<ul>
<li>Allow using LIST, Create Containers using custom-llm-provider - <a href="https://github.com/BerriAI/litellm/pull/17740" target="_blank" rel="noopener noreferrer">PR #17740</a></li>
<li>Add new container API file management + UI Interface - <a href="https://github.com/BerriAI/litellm/pull/17745" target="_blank" rel="noopener noreferrer">PR #17745</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/rerank">Rerank API</a></strong>
<ul>
<li>Add support for forwarding client headers in /rerank endpoint - <a href="https://github.com/BerriAI/litellm/pull/17873" target="_blank" rel="noopener noreferrer">PR #17873</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/files_endpoints">Files API</a></strong>
<ul>
<li>Add support for expires_after param in Files endpoint - <a href="https://github.com/BerriAI/litellm/pull/17860" target="_blank" rel="noopener noreferrer">PR #17860</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/videos">Video API</a></strong>
<ul>
<li>Use litellm params for all videos APIs - <a href="https://github.com/BerriAI/litellm/pull/17732" target="_blank" rel="noopener noreferrer">PR #17732</a></li>
<li>Respect videos content db creds - <a href="https://github.com/BerriAI/litellm/pull/17771" target="_blank" rel="noopener noreferrer">PR #17771</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/embedding">Embeddings API</a></strong>
<ul>
<li>Fix handling token array input decoding for embeddings - <a href="https://github.com/BerriAI/litellm/pull/17468" target="_blank" rel="noopener noreferrer">PR #17468</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/completion/input">Chat Completions API</a></strong>
<ul>
<li>Add v0 target storage support - store files in Azure AI storage and use with chat completions API - <a href="https://github.com/BerriAI/litellm/pull/17758" target="_blank" rel="noopener noreferrer">PR #17758</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/gemini">generateContent API</a></strong>
<ul>
<li>Support model names with slashes on Gemini generateContent endpoints - <a href="https://github.com/BerriAI/litellm/pull/17743" target="_blank" rel="noopener noreferrer">PR #17743</a></li>
</ul>
</li>
<li><strong>General</strong>
<ul>
<li>Use audio content for caching - <a href="https://github.com/BerriAI/litellm/pull/17651" target="_blank" rel="noopener noreferrer">PR #17651</a></li>
<li>Return 403 exception when calling GET responses API - <a href="https://github.com/BerriAI/litellm/pull/17629" target="_blank" rel="noopener noreferrer">PR #17629</a></li>
<li>Add nested field removal support to additional_drop_params - <a href="https://github.com/BerriAI/litellm/pull/17711" target="_blank" rel="noopener noreferrer">PR #17711</a></li>
<li>Async post_call_streaming_iterator_hook now properly iterates async generators - <a href="https://github.com/BerriAI/litellm/pull/17626" target="_blank" rel="noopener noreferrer">PR #17626</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs">Bugs<a href="https://docs.litellm.ai/release_notes/v1-80-10#bugs" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>General</strong>
<ul>
<li>Fix handle string content in is_cached_message - <a href="https://github.com/BerriAI/litellm/pull/17853" target="_blank" rel="noopener noreferrer">PR #17853</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-80-10#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-80-10#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li><strong>UI Settings</strong>
<ul>
<li>Add Get and Update Backend Routes for UI Settings - <a href="https://github.com/BerriAI/litellm/pull/17689" target="_blank" rel="noopener noreferrer">PR #17689</a></li>
<li>UI Settings page implementation - <a href="https://github.com/BerriAI/litellm/pull/17697" target="_blank" rel="noopener noreferrer">PR #17697</a></li>
<li>Ensure Model Page honors UI Settings - <a href="https://github.com/BerriAI/litellm/pull/17804" target="_blank" rel="noopener noreferrer">PR #17804</a></li>
<li>Add All Proxy Models to Default User Settings - <a href="https://github.com/BerriAI/litellm/pull/17902" target="_blank" rel="noopener noreferrer">PR #17902</a></li>
</ul>
</li>
<li><strong>Agent &amp; Usage UI</strong>
<ul>
<li>Daily Agent Usage Backend - <a href="https://github.com/BerriAI/litellm/pull/17781" target="_blank" rel="noopener noreferrer">PR #17781</a></li>
<li>Agent Usage UI - <a href="https://github.com/BerriAI/litellm/pull/17797" target="_blank" rel="noopener noreferrer">PR #17797</a></li>
<li>Add agent cost tracking on UI - <a href="https://github.com/BerriAI/litellm/pull/17899" target="_blank" rel="noopener noreferrer">PR #17899</a></li>
<li>New Badge for Agent Usage - <a href="https://github.com/BerriAI/litellm/pull/17883" target="_blank" rel="noopener noreferrer">PR #17883</a></li>
<li>Usage Entity labels for filtering - <a href="https://github.com/BerriAI/litellm/pull/17896" target="_blank" rel="noopener noreferrer">PR #17896</a></li>
<li>Agent Usage Page minor fixes - <a href="https://github.com/BerriAI/litellm/pull/17901" target="_blank" rel="noopener noreferrer">PR #17901</a></li>
<li>Usage Page View Select component - <a href="https://github.com/BerriAI/litellm/pull/17854" target="_blank" rel="noopener noreferrer">PR #17854</a></li>
<li>Usage Page Components refactor - <a href="https://github.com/BerriAI/litellm/pull/17848" target="_blank" rel="noopener noreferrer">PR #17848</a></li>
</ul>
</li>
<li><strong>Logs &amp; Spend</strong>
<ul>
<li>Enhanced spend analytics in logs view - <a href="https://github.com/BerriAI/litellm/pull/17623" target="_blank" rel="noopener noreferrer">PR #17623</a></li>
<li>Add user info delete modal for user management - <a href="https://github.com/BerriAI/litellm/pull/17625" target="_blank" rel="noopener noreferrer">PR #17625</a></li>
<li>Show request and response details in logs view - <a href="https://github.com/BerriAI/litellm/pull/17928" target="_blank" rel="noopener noreferrer">PR #17928</a></li>
</ul>
</li>
<li><strong>Virtual Keys</strong>
<ul>
<li>Fix x-litellm-key-spend header update - <a href="https://github.com/BerriAI/litellm/pull/17864" target="_blank" rel="noopener noreferrer">PR #17864</a></li>
</ul>
</li>
<li><strong>Models &amp; Endpoints</strong>
<ul>
<li>Model Hub Useful Links Rearrange - <a href="https://github.com/BerriAI/litellm/pull/17859" target="_blank" rel="noopener noreferrer">PR #17859</a></li>
<li>Create Team Model Dropdown honors Organization's Models - <a href="https://github.com/BerriAI/litellm/pull/17834" target="_blank" rel="noopener noreferrer">PR #17834</a></li>
</ul>
</li>
<li><strong>SSO &amp; Auth</strong>
<ul>
<li>Allow upserting user role when SSO provider role changes - <a href="https://github.com/BerriAI/litellm/pull/17754" target="_blank" rel="noopener noreferrer">PR #17754</a></li>
<li>Allow fetching role from generic SSO provider (Keycloak) - <a href="https://github.com/BerriAI/litellm/pull/17787" target="_blank" rel="noopener noreferrer">PR #17787</a></li>
<li>JWT Auth - allow selecting team_id from request header - <a href="https://github.com/BerriAI/litellm/pull/17884" target="_blank" rel="noopener noreferrer">PR #17884</a></li>
<li>Remove SSO Config Values from Config Table on SSO Update - <a href="https://github.com/BerriAI/litellm/pull/17668" target="_blank" rel="noopener noreferrer">PR #17668</a></li>
</ul>
</li>
<li><strong>Teams</strong>
<ul>
<li>Attach team to org table - <a href="https://github.com/BerriAI/litellm/pull/17832" target="_blank" rel="noopener noreferrer">PR #17832</a></li>
<li>Expose the team alias when authenticating - <a href="https://github.com/BerriAI/litellm/pull/17725" target="_blank" rel="noopener noreferrer">PR #17725</a></li>
</ul>
</li>
<li><strong>MCP Server Management</strong>
<ul>
<li>Add extra_headers and allowed_tools to UpdateMCPServerRequest - <a href="https://github.com/BerriAI/litellm/pull/17940" target="_blank" rel="noopener noreferrer">PR #17940</a></li>
</ul>
</li>
<li><strong>Notifications</strong>
<ul>
<li>Show progress and pause on hover for Notifications - <a href="https://github.com/BerriAI/litellm/pull/17942" target="_blank" rel="noopener noreferrer">PR #17942</a></li>
</ul>
</li>
<li><strong>General</strong>
<ul>
<li>Allow Root Path to Redirect when Docs not on Root Path - <a href="https://github.com/BerriAI/litellm/pull/16843" target="_blank" rel="noopener noreferrer">PR #16843</a></li>
<li>Show UI version number on top left near logo - <a href="https://github.com/BerriAI/litellm/pull/17891" target="_blank" rel="noopener noreferrer">PR #17891</a></li>
<li>Re-organize left navigation with correct categories and agents on root - <a href="https://github.com/BerriAI/litellm/pull/17890" target="_blank" rel="noopener noreferrer">PR #17890</a></li>
<li>UI Playground - allow custom model names in model selector dropdown - <a href="https://github.com/BerriAI/litellm/pull/17892" target="_blank" rel="noopener noreferrer">PR #17892</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs-1">Bugs<a href="https://docs.litellm.ai/release_notes/v1-80-10#bugs-1" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>UI Fixes</strong>
<ul>
<li>Fix links + old login page deprecation message - <a href="https://github.com/BerriAI/litellm/pull/17624" target="_blank" rel="noopener noreferrer">PR #17624</a></li>
<li>Filtering for Chat UI Endpoint Selector - <a href="https://github.com/BerriAI/litellm/pull/17567" target="_blank" rel="noopener noreferrer">PR #17567</a></li>
<li>Race Condition Handling in SCIM v2 - <a href="https://github.com/BerriAI/litellm/pull/17513" target="_blank" rel="noopener noreferrer">PR #17513</a></li>
<li>Make /litellm_model_cost_map public - <a href="https://github.com/BerriAI/litellm/pull/16795" target="_blank" rel="noopener noreferrer">PR #16795</a></li>
<li>Custom Callback on UI - <a href="https://github.com/BerriAI/litellm/pull/17522" target="_blank" rel="noopener noreferrer">PR #17522</a></li>
<li>Add User Writable Directory to Non Root Docker for Logo - <a href="https://github.com/BerriAI/litellm/pull/17180" target="_blank" rel="noopener noreferrer">PR #17180</a></li>
<li>Swap URL Input and Display Name inputs - <a href="https://github.com/BerriAI/litellm/pull/17682" target="_blank" rel="noopener noreferrer">PR #17682</a></li>
<li>Change deprecation banner to only show on /sso/key/generate - <a href="https://github.com/BerriAI/litellm/pull/17681" target="_blank" rel="noopener noreferrer">PR #17681</a></li>
<li>Change credential encryption to only affect db credentials - <a href="https://github.com/BerriAI/litellm/pull/17741" target="_blank" rel="noopener noreferrer">PR #17741</a></li>
</ul>
</li>
<li><strong>Auth &amp; Routes</strong>
<ul>
<li>Return 403 instead of 503 for unauthorized routes - <a href="https://github.com/BerriAI/litellm/pull/17723" target="_blank" rel="noopener noreferrer">PR #17723</a></li>
<li>AI Gateway Auth - allow using wildcard patterns for public routes - <a href="https://github.com/BerriAI/litellm/pull/17686" target="_blank" rel="noopener noreferrer">PR #17686</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="ai-integrations">AI Integrations<a href="https://docs.litellm.ai/release_notes/v1-80-10#ai-integrations" class="hash-link" aria-label="Direct link to AI Integrations" title="Direct link to AI Integrations">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="new-integrations-4-new-integrations">New Integrations (4 new integrations)<a href="https://docs.litellm.ai/release_notes/v1-80-10#new-integrations-4-new-integrations" class="hash-link" aria-label="Direct link to New Integrations (4 new integrations)" title="Direct link to New Integrations (4 new integrations)">​</a></h3>
<table><thead><tr><th>Integration</th><th>Type</th><th>Description</th></tr></thead><tbody><tr><td><a href="https://docs.litellm.ai/docs/proxy/logging#sumologic">SumoLogic</a></td><td>Logging</td><td>Native webhook integration for SumoLogic - <a href="https://github.com/BerriAI/litellm/pull/17630" target="_blank" rel="noopener noreferrer">PR #17630</a></td></tr><tr><td><a href="https://docs.litellm.ai/docs/proxy/arize_phoenix_prompts">Arize Phoenix</a></td><td>Prompt Management</td><td>Arize Phoenix OSS prompt management integration - <a href="https://github.com/BerriAI/litellm/pull/17750" target="_blank" rel="noopener noreferrer">PR #17750</a></td></tr><tr><td><a href="https://docs.litellm.ai/docs/proxy/email">Sendgrid</a></td><td>Email</td><td>Sendgrid email notifications integration - <a href="https://github.com/BerriAI/litellm/pull/17775" target="_blank" rel="noopener noreferrer">PR #17775</a></td></tr><tr><td><a href="https://docs.litellm.ai/docs/proxy/guardrails/onyx_security">Onyx</a></td><td>Guardrails</td><td>Onyx guardrail hooks integration - <a href="https://github.com/BerriAI/litellm/pull/16591" target="_blank" rel="noopener noreferrer">PR #16591</a></td></tr></tbody></table>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="logging">Logging<a href="https://docs.litellm.ai/release_notes/v1-80-10#logging" class="hash-link" aria-label="Direct link to Logging" title="Direct link to Logging">​</a></h3>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging#langfuse">Langfuse</a></strong>
<ul>
<li>Propagate Langfuse trace_id - <a href="https://github.com/BerriAI/litellm/pull/17669" target="_blank" rel="noopener noreferrer">PR #17669</a></li>
<li>Prefer standard trace id for Langfuse logging - <a href="https://github.com/BerriAI/litellm/pull/17791" target="_blank" rel="noopener noreferrer">PR #17791</a></li>
<li>Move query params to create_pass_through_route call in Langfuse passthrough - <a href="https://github.com/BerriAI/litellm/pull/17660" target="_blank" rel="noopener noreferrer">PR #17660</a></li>
<li>Add support for custom masking function - <a href="https://github.com/BerriAI/litellm/pull/17826" target="_blank" rel="noopener noreferrer">PR #17826</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging#prometheus">Prometheus</a></strong>
<ul>
<li>Add 'exception_status' to prometheus logger - <a href="https://github.com/BerriAI/litellm/pull/17847" target="_blank" rel="noopener noreferrer">PR #17847</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging#otel">OpenTelemetry</a></strong>
<ul>
<li>Add latency metrics (TTFT, TPOT, Total Generation Time) to OTEL payload - <a href="https://github.com/BerriAI/litellm/pull/17888" target="_blank" rel="noopener noreferrer">PR #17888</a></li>
</ul>
</li>
<li><strong>General</strong>
<ul>
<li>Add polling via cache feature for async logging - <a href="https://github.com/BerriAI/litellm/pull/16862" target="_blank" rel="noopener noreferrer">PR #16862</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails">Guardrails<a href="https://docs.litellm.ai/release_notes/v1-80-10#guardrails" class="hash-link" aria-label="Direct link to Guardrails" title="Direct link to Guardrails">​</a></h3>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/hiddenlayer">HiddenLayer</a></strong>
<ul>
<li>Add HiddenLayer Guardrail Hooks - <a href="https://github.com/BerriAI/litellm/pull/17728" target="_blank" rel="noopener noreferrer">PR #17728</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/pillar_security">Pillar Security</a></strong>
<ul>
<li>Add opt-in evidence results for Pillar Security guardrail during monitoring - <a href="https://github.com/BerriAI/litellm/pull/17812" target="_blank" rel="noopener noreferrer">PR #17812</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/panw_prisma_airs">PANW Prisma AIRS</a></strong>
<ul>
<li>Add configurable fail-open, timeout, and app_user tracking - <a href="https://github.com/BerriAI/litellm/pull/17785" target="_blank" rel="noopener noreferrer">PR #17785</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/pii_masking_v2">Presidio</a></strong>
<ul>
<li>Add support for configurable confidence score thresholds and scope in Presidio PII masking - <a href="https://github.com/BerriAI/litellm/pull/17817" target="_blank" rel="noopener noreferrer">PR #17817</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/litellm_content_filter">LiteLLM Content Filter</a></strong>
<ul>
<li>Mask all regex pattern matches, not just first - <a href="https://github.com/BerriAI/litellm/pull/17727" target="_blank" rel="noopener noreferrer">PR #17727</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/secret_detection">Regex Guardrails</a></strong>
<ul>
<li>Add enhanced regex pattern matching for guardrails - <a href="https://github.com/BerriAI/litellm/pull/17915" target="_blank" rel="noopener noreferrer">PR #17915</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/grayswan">Gray Swan Guardrail</a></strong>
<ul>
<li>Add passthrough mode for model response - <a href="https://github.com/BerriAI/litellm/pull/17102" target="_blank" rel="noopener noreferrer">PR #17102</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="prompt-management">Prompt Management<a href="https://docs.litellm.ai/release_notes/v1-80-10#prompt-management" class="hash-link" aria-label="Direct link to Prompt Management" title="Direct link to Prompt Management">​</a></h3>
<ul>
<li><strong>General</strong>
<ul>
<li>New API for integrating prompt management providers - <a href="https://github.com/BerriAI/litellm/pull/17829" target="_blank" rel="noopener noreferrer">PR #17829</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-80-10#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li><strong>Service Tier Pricing</strong> - Extract service_tier from response/usage for OpenAI flex pricing - <a href="https://github.com/BerriAI/litellm/pull/17748" target="_blank" rel="noopener noreferrer">PR #17748</a></li>
<li><strong>Agent Cost Tracking</strong> - Track agent_id in SpendLogs - <a href="https://github.com/BerriAI/litellm/pull/17795" target="_blank" rel="noopener noreferrer">PR #17795</a></li>
<li><strong>Tag Activity</strong> - Deduplicate /tag/daily/activity metadata - <a href="https://github.com/BerriAI/litellm/pull/16764" target="_blank" rel="noopener noreferrer">PR #16764</a></li>
<li><strong>Rate Limiting</strong> - Dynamic Rate Limiter - allow specifying ttl for in memory cache - <a href="https://github.com/BerriAI/litellm/pull/17679" target="_blank" rel="noopener noreferrer">PR #17679</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-gateway">MCP Gateway<a href="https://docs.litellm.ai/release_notes/v1-80-10#mcp-gateway" class="hash-link" aria-label="Direct link to MCP Gateway" title="Direct link to MCP Gateway">​</a></h2>
<ul>
<li><strong>Chat Completions Integration</strong> - Add support for using MCPs on /chat/completions - <a href="https://github.com/BerriAI/litellm/pull/17747" target="_blank" rel="noopener noreferrer">PR #17747</a></li>
<li><strong>UI Session Permissions</strong> - Fix UI session MCP permissions across real teams - <a href="https://github.com/BerriAI/litellm/pull/17620" target="_blank" rel="noopener noreferrer">PR #17620</a></li>
<li><strong>OAuth Callback</strong> - Fix MCP OAuth callback routing and URL handling - <a href="https://github.com/BerriAI/litellm/pull/17789" target="_blank" rel="noopener noreferrer">PR #17789</a></li>
<li><strong>Tool Name Prefix</strong> - Fix MCP tool name prefix - <a href="https://github.com/BerriAI/litellm/pull/17908" target="_blank" rel="noopener noreferrer">PR #17908</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="agent-gateway-a2a">Agent Gateway (A2A)<a href="https://docs.litellm.ai/release_notes/v1-80-10#agent-gateway-a2a" class="hash-link" aria-label="Direct link to Agent Gateway (A2A)" title="Direct link to Agent Gateway (A2A)">​</a></h2>
<ul>
<li><strong>Cost Per Query</strong> - Add cost per query for agent invocations - <a href="https://github.com/BerriAI/litellm/pull/17774" target="_blank" rel="noopener noreferrer">PR #17774</a></li>
<li><strong>Token Counting</strong> - Add token counting non streaming + streaming - <a href="https://github.com/BerriAI/litellm/pull/17779" target="_blank" rel="noopener noreferrer">PR #17779</a></li>
<li><strong>Cost Per Token</strong> - Add cost per token pricing for A2A - <a href="https://github.com/BerriAI/litellm/pull/17780" target="_blank" rel="noopener noreferrer">PR #17780</a></li>
<li><strong>LangGraph Provider</strong> - Add LangGraph provider for Agent Gateway - <a href="https://github.com/BerriAI/litellm/pull/17783" target="_blank" rel="noopener noreferrer">PR #17783</a></li>
<li><strong>Bedrock &amp; LangGraph Agents</strong> - Allow using Bedrock AgentCore, LangGraph agents with A2A Gateway - <a href="https://github.com/BerriAI/litellm/pull/17786" target="_blank" rel="noopener noreferrer">PR #17786</a></li>
<li><strong>Agent Management</strong> - Allow adding LangGraph, Bedrock Agent Core agents - <a href="https://github.com/BerriAI/litellm/pull/17802" target="_blank" rel="noopener noreferrer">PR #17802</a></li>
<li><strong>Azure Foundry Agents</strong> - Add Azure AI Foundry Agents support - <a href="https://github.com/BerriAI/litellm/pull/17845" target="_blank" rel="noopener noreferrer">PR #17845</a></li>
<li><strong>Azure Foundry UI</strong> - Allow adding Azure Foundry Agents on UI - <a href="https://github.com/BerriAI/litellm/pull/17909" target="_blank" rel="noopener noreferrer">PR #17909</a></li>
<li><strong>Azure Foundry Fixes</strong> - Ensure Azure Foundry agents work correctly - <a href="https://github.com/BerriAI/litellm/pull/17943" target="_blank" rel="noopener noreferrer">PR #17943</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements">Performance / Loadbalancing / Reliability improvements<a href="https://docs.litellm.ai/release_notes/v1-80-10#performance--loadbalancing--reliability-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements" title="Direct link to Performance / Loadbalancing / Reliability improvements">​</a></h2>
<ul>
<li><strong>Memory Leak Fix</strong> - Cut memory leak in half - <a href="https://github.com/BerriAI/litellm/pull/17784" target="_blank" rel="noopener noreferrer">PR #17784</a></li>
<li><strong>Spend Logs Memory</strong> - Reduce memory accumulation of spend_logs - <a href="https://github.com/BerriAI/litellm/pull/17742" target="_blank" rel="noopener noreferrer">PR #17742</a></li>
<li><strong>Router Optimization</strong> - Replace time.perf_counter() with time.time() - <a href="https://github.com/BerriAI/litellm/pull/17881" target="_blank" rel="noopener noreferrer">PR #17881</a></li>
<li><strong>Filter Internal Params</strong> - Filter internal params in fallback code - <a href="https://github.com/BerriAI/litellm/pull/17941" target="_blank" rel="noopener noreferrer">PR #17941</a></li>
<li><strong>Gunicorn Suggestion</strong> - Suggest Gunicorn instead of uvicorn when using max_requests_before_restart - <a href="https://github.com/BerriAI/litellm/pull/17788" target="_blank" rel="noopener noreferrer">PR #17788</a></li>
<li><strong>Pydantic Warnings</strong> - Mitigate PydanticDeprecatedSince20 warnings - <a href="https://github.com/BerriAI/litellm/pull/17657" target="_blank" rel="noopener noreferrer">PR #17657</a></li>
<li><strong>Python 3.14 Support</strong> - Add Python 3.14 support via grpcio version constraints - <a href="https://github.com/BerriAI/litellm/pull/17666" target="_blank" rel="noopener noreferrer">PR #17666</a></li>
<li><strong>OpenAI Package</strong> - Bump openai package to 2.9.0 - <a href="https://github.com/BerriAI/litellm/pull/17818" target="_blank" rel="noopener noreferrer">PR #17818</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates">Documentation Updates<a href="https://docs.litellm.ai/release_notes/v1-80-10#documentation-updates" class="hash-link" aria-label="Direct link to Documentation Updates" title="Direct link to Documentation Updates">​</a></h2>
<ul>
<li><strong>Contributing</strong> - Update clone instructions to recommend forking first - <a href="https://github.com/BerriAI/litellm/pull/17637" target="_blank" rel="noopener noreferrer">PR #17637</a></li>
<li><strong>Getting Started</strong> - Improve Getting Started page and SDK documentation structure - <a href="https://github.com/BerriAI/litellm/pull/17614" target="_blank" rel="noopener noreferrer">PR #17614</a></li>
<li><strong>JSON Mode</strong> - Make it clearer how to get Pydantic model output - <a href="https://github.com/BerriAI/litellm/pull/17671" target="_blank" rel="noopener noreferrer">PR #17671</a></li>
<li><strong>drop_params</strong> - Update litellm docs for drop_params - <a href="https://github.com/BerriAI/litellm/pull/17658" target="_blank" rel="noopener noreferrer">PR #17658</a></li>
<li><strong>Environment Variables</strong> - Document missing environment variables and fix incorrect types - <a href="https://github.com/BerriAI/litellm/pull/17649" target="_blank" rel="noopener noreferrer">PR #17649</a></li>
<li><strong>SumoLogic</strong> - Add SumoLogic integration documentation - <a href="https://github.com/BerriAI/litellm/pull/17647" target="_blank" rel="noopener noreferrer">PR #17647</a></li>
<li><strong>SAP Gen AI</strong> - Add SAP Gen AI provider documentation - <a href="https://github.com/BerriAI/litellm/pull/17667" target="_blank" rel="noopener noreferrer">PR #17667</a></li>
<li><strong>Authentication</strong> - Add Note for Authentication - <a href="https://github.com/BerriAI/litellm/pull/17733" target="_blank" rel="noopener noreferrer">PR #17733</a></li>
<li><strong>Known Issues</strong> - Adding known issues to 1.80.5-stable docs - <a href="https://github.com/BerriAI/litellm/pull/17738" target="_blank" rel="noopener noreferrer">PR #17738</a></li>
<li><strong>Supported Endpoints</strong> - Fix Supported Endpoints page - <a href="https://github.com/BerriAI/litellm/pull/17710" target="_blank" rel="noopener noreferrer">PR #17710</a></li>
<li><strong>Token Count</strong> - Document token count endpoint - <a href="https://github.com/BerriAI/litellm/pull/17772" target="_blank" rel="noopener noreferrer">PR #17772</a></li>
<li><strong>Overview</strong> - Made litellm proxy and SDK difference cleaner in overview with a table - <a href="https://github.com/BerriAI/litellm/pull/17790" target="_blank" rel="noopener noreferrer">PR #17790</a></li>
<li><strong>Containers API</strong> - Add docs for containers files API + code interpreter on LiteLLM - <a href="https://github.com/BerriAI/litellm/pull/17749" target="_blank" rel="noopener noreferrer">PR #17749</a></li>
<li><strong>Target Storage</strong> - Add documentation for target storage - <a href="https://github.com/BerriAI/litellm/pull/17882" target="_blank" rel="noopener noreferrer">PR #17882</a></li>
<li><strong>Agent Usage</strong> - Agent Usage documentation - <a href="https://github.com/BerriAI/litellm/pull/17931" target="_blank" rel="noopener noreferrer">PR #17931</a>, <a href="https://github.com/BerriAI/litellm/pull/17932" target="_blank" rel="noopener noreferrer">PR #17932</a>, <a href="https://github.com/BerriAI/litellm/pull/17934" target="_blank" rel="noopener noreferrer">PR #17934</a></li>
<li><strong>Cursor Integration</strong> - Cursor Integration documentation - <a href="https://github.com/BerriAI/litellm/pull/17855" target="_blank" rel="noopener noreferrer">PR #17855</a>, <a href="https://github.com/BerriAI/litellm/pull/17939" target="_blank" rel="noopener noreferrer">PR #17939</a></li>
<li><strong>A2A Cost Tracking</strong> - A2A cost tracking docs - <a href="https://github.com/BerriAI/litellm/pull/17913" target="_blank" rel="noopener noreferrer">PR #17913</a></li>
<li><strong>Azure Search</strong> - Update azure search docs - <a href="https://github.com/BerriAI/litellm/pull/17726" target="_blank" rel="noopener noreferrer">PR #17726</a></li>
<li><strong>Milvus Client</strong> - Fix milvus client docs - <a href="https://github.com/BerriAI/litellm/pull/17736" target="_blank" rel="noopener noreferrer">PR #17736</a></li>
<li><strong>Streaming Logging</strong> - Remove streaming logging doc - <a href="https://github.com/BerriAI/litellm/pull/17739" target="_blank" rel="noopener noreferrer">PR #17739</a></li>
<li><strong>Integration Docs</strong> - Update integration docs location - <a href="https://github.com/BerriAI/litellm/pull/17644" target="_blank" rel="noopener noreferrer">PR #17644</a></li>
<li><strong>Links</strong> - Updated docs links for mistral and anthropic - <a href="https://github.com/BerriAI/litellm/pull/17852" target="_blank" rel="noopener noreferrer">PR #17852</a></li>
<li><strong>Community</strong> - Add community doc link - <a href="https://github.com/BerriAI/litellm/pull/17734" target="_blank" rel="noopener noreferrer">PR #17734</a></li>
<li><strong>Pricing</strong> - Update pricing for global.anthropic.claude-haiku-4-5-20251001-v1:0 - <a href="https://github.com/BerriAI/litellm/pull/17703" target="_blank" rel="noopener noreferrer">PR #17703</a></li>
<li><strong>gpt-image-1-mini</strong> - Correct model type for gpt-image-1-mini - <a href="https://github.com/BerriAI/litellm/pull/17635" target="_blank" rel="noopener noreferrer">PR #17635</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="infrastructure--deployment">Infrastructure / Deployment<a href="https://docs.litellm.ai/release_notes/v1-80-10#infrastructure--deployment" class="hash-link" aria-label="Direct link to Infrastructure / Deployment" title="Direct link to Infrastructure / Deployment">​</a></h2>
<ul>
<li><strong>Docker</strong> - Use python instead of wget for healthcheck in docker-compose.yml - <a href="https://github.com/BerriAI/litellm/pull/17646" target="_blank" rel="noopener noreferrer">PR #17646</a></li>
<li><strong>Helm Chart</strong> - Add extraResources support for Helm chart deployments - <a href="https://github.com/BerriAI/litellm/pull/17627" target="_blank" rel="noopener noreferrer">PR #17627</a></li>
<li><strong>Helm Versioning</strong> - Add semver prerelease suffix to helm chart versions - <a href="https://github.com/BerriAI/litellm/pull/17678" target="_blank" rel="noopener noreferrer">PR #17678</a></li>
<li><strong>Database Schema</strong> - Add storage_backend and storage_url columns to schema.prisma for target storage feature - <a href="https://github.com/BerriAI/litellm/pull/17936" target="_blank" rel="noopener noreferrer">PR #17936</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-80-10#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@xianzongxie-stripe made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16862" target="_blank" rel="noopener noreferrer">PR #16862</a></li>
<li>@krisxia0506 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17637" target="_blank" rel="noopener noreferrer">PR #17637</a></li>
<li>@chetanchoudhary-sumo made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17630" target="_blank" rel="noopener noreferrer">PR #17630</a></li>
<li>@kevinmarx made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17632" target="_blank" rel="noopener noreferrer">PR #17632</a></li>
<li>@expruc made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17627" target="_blank" rel="noopener noreferrer">PR #17627</a></li>
<li>@rcII made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17626" target="_blank" rel="noopener noreferrer">PR #17626</a></li>
<li>@tamirkiviti13 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16591" target="_blank" rel="noopener noreferrer">PR #16591</a></li>
<li>@Eric84626 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17629" target="_blank" rel="noopener noreferrer">PR #17629</a></li>
<li>@vasilisazayka made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16053" target="_blank" rel="noopener noreferrer">PR #16053</a></li>
<li>@juliettech13 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17663" target="_blank" rel="noopener noreferrer">PR #17663</a></li>
<li>@jason-nance made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17660" target="_blank" rel="noopener noreferrer">PR #17660</a></li>
<li>@yisding made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17671" target="_blank" rel="noopener noreferrer">PR #17671</a></li>
<li>@emilsvennesson made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17656" target="_blank" rel="noopener noreferrer">PR #17656</a></li>
<li>@kumekay made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17646" target="_blank" rel="noopener noreferrer">PR #17646</a></li>
<li>@chenzhaofei01 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17584" target="_blank" rel="noopener noreferrer">PR #17584</a></li>
<li>@shivamrawat1 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17733" target="_blank" rel="noopener noreferrer">PR #17733</a></li>
<li>@ephrimstanley made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17723" target="_blank" rel="noopener noreferrer">PR #17723</a></li>
<li>@hwittenborn made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17743" target="_blank" rel="noopener noreferrer">PR #17743</a></li>
<li>@peterkc made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17727" target="_blank" rel="noopener noreferrer">PR #17727</a></li>
<li>@saisurya237 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17725" target="_blank" rel="noopener noreferrer">PR #17725</a></li>
<li>@Ashton-Sidhu made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17728" target="_blank" rel="noopener noreferrer">PR #17728</a></li>
<li>@CyrusTC made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17810" target="_blank" rel="noopener noreferrer">PR #17810</a></li>
<li>@jichmi made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17703" target="_blank" rel="noopener noreferrer">PR #17703</a></li>
<li>@ryan-crabbe made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17852" target="_blank" rel="noopener noreferrer">PR #17852</a></li>
<li>@nlineback made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17851" target="_blank" rel="noopener noreferrer">PR #17851</a></li>
<li>@butnarurazvan made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17468" target="_blank" rel="noopener noreferrer">PR #17468</a></li>
<li>@yoshi-p27 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17915" target="_blank" rel="noopener noreferrer">PR #17915</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog">Full Changelog<a href="https://docs.litellm.ai/release_notes/v1-80-10#full-changelog" class="hash-link" aria-label="Direct link to Full Changelog" title="Direct link to Full Changelog">​</a></h2>
<p><strong><a href="https://github.com/BerriAI/litellm/compare/v1.80.8.rc.1...v1.80.10" target="_blank" rel="noopener noreferrer">View complete changelog on GitHub</a></strong></p>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[v1.80.8-stable - Introducing A2A Agent Gateway]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-80-8</id>
        <link href="https://docs.litellm.ai/release_notes/v1-80-8"/>
        <updated>2025-12-06T10:00:00.000Z</updated>
        <summary type="html"><![CDATA[Deploy this version]]></summary>
        <content type="html"><![CDATA[<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-80-8#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">docker.litellm.ai/berriai/litellm:v1.80.8-stable</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.80.8</span><br></span></code></pre></div></div></div></div></div>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-80-8#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<ul>
<li><strong>Agent Gateway (A2A)</strong> - <a href="https://docs.litellm.ai/docs/a2a">Invoke agents through the AI Gateway with request/response logging and access controls</a></li>
<li><strong>Guardrails API v2</strong> - <a href="https://docs.litellm.ai/docs/adding_provider/generic_guardrail_api">Generic Guardrail API with streaming support, structured messages, and tool call checks</a></li>
<li><strong>Customer (End User) Usage UI</strong> - <a href="https://docs.litellm.ai/docs/proxy/customer_usage">Track and visualize end-user spend directly in the dashboard</a></li>
<li><strong>vLLM Batch + Files API</strong> - <a href="https://docs.litellm.ai/docs/batches">Support for batch and files API with vLLM deployments</a></li>
<li><strong>Dynamic Rate Limiting on Teams</strong> - <a href="https://docs.litellm.ai/docs/proxy/team_budgets">Enable dynamic rate limits and priority reservation on team-level</a></li>
<li><strong>Google Cloud Chirp3 HD</strong> - <a href="https://docs.litellm.ai/docs/text_to_speech">New text-to-speech provider with Chirp3 HD voices</a></li>
</ul>
<hr>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="agent-gateway-a2a">Agent Gateway (A2A)<a href="https://docs.litellm.ai/release_notes/v1-80-8#agent-gateway-a2a" class="hash-link" aria-label="Direct link to Agent Gateway (A2A)" title="Direct link to Agent Gateway (A2A)">​</a></h3>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAGCAIAAAB1kpiRAAAACXBIWXMAAAsTAAALEwEAmpwYAAAAt0lEQVR4nGPg5JDi5JBiZ5OQEFeXl9aUl9IQF1NnZxVnYxLmYBVnYGMRZ2UWY2eT4OeSktRzlbQI4uOS5uJTENGw5xXRYBCV0pJRMeHmleeUMrKYusd71QUB00AeUXUduxAJRQsGIREVMSltkBVShtZ9G+xm7Re0DOPmlZMVUxQWUGZgY5FgZRbnYJNiZxQR13WRtw3jYBXn4pHnF1Ti4ZZl4GCXgiMuFlFuZmEOdpBL2Vgl2NkkAShQFaYvCJPVAAAAAElFTkSuQmCC&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="381"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/a2a_gateway.6f8c840.640.png" srcset="/assets/ideal-img/a2a_gateway.6f8c840.640.png 640w,/assets/ideal-img/a2a_gateway.2ce6ee2.1344.png 1344w" width="640" height="381"></noscript></div>
<br>
<p>This release introduces <strong>A2A Agent Gateway</strong> for LiteLLM, allowing you to invoke and manage A2A agents with the same controls you have for LLM APIs.</p>
<p>As a <strong>LiteLLM Gateway Admin</strong>, you can now do the following:</p>
<ul>
<li><strong>Request/Response Logging</strong> - Every agent invocation is logged to the Logs page with full request and response tracking.</li>
<li><strong>Access Control</strong> - Control which Team/Key can access which agents.</li>
</ul>
<p>As a developer, you can continue using the A2A SDK, all you need to do is point you <code>A2AClient</code> to the LiteLLM proxy URL and your API key.</p>
<p><strong>Works with the A2A SDK:</strong></p>
<div class="language-python codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-python codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token keyword" style="color:#00009f">from</span><span class="token plain"> a2a</span><span class="token punctuation" style="color:#393A34">.</span><span class="token plain">client </span><span class="token keyword" style="color:#00009f">import</span><span class="token plain"> A2AClient</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain" style="display:inline-block"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">client </span><span class="token operator" style="color:#393A34">=</span><span class="token plain"> A2AClient</span><span class="token punctuation" style="color:#393A34">(</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">    base_url</span><span class="token operator" style="color:#393A34">=</span><span class="token string" style="color:#e3116c">"http://localhost:4000"</span><span class="token punctuation" style="color:#393A34">,</span><span class="token plain">  </span><span class="token comment" style="color:#999988;font-style:italic"># Your LiteLLM proxy</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">    api_key</span><span class="token operator" style="color:#393A34">=</span><span class="token string" style="color:#e3116c">"sk-1234"</span><span class="token plain">                   </span><span class="token comment" style="color:#999988;font-style:italic"># LiteLLM API key</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain"></span><span class="token punctuation" style="color:#393A34">)</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain" style="display:inline-block"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">response </span><span class="token operator" style="color:#393A34">=</span><span class="token plain"> client</span><span class="token punctuation" style="color:#393A34">.</span><span class="token plain">send_message</span><span class="token punctuation" style="color:#393A34">(</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">    agent_id</span><span class="token operator" style="color:#393A34">=</span><span class="token string" style="color:#e3116c">"my-agent"</span><span class="token punctuation" style="color:#393A34">,</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">    message</span><span class="token operator" style="color:#393A34">=</span><span class="token string" style="color:#e3116c">"What's the status of my order?"</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain"></span><span class="token punctuation" style="color:#393A34">)</span><br></span></code></pre></div></div>
<p>Get started with Agent Gateway here: <a href="https://docs.litellm.ai/docs/a2a">Agent Gateway Documentation</a></p>
<hr>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="customer-end-user-usage-ui">Customer (End User) Usage UI<a href="https://docs.litellm.ai/release_notes/v1-80-8#customer-end-user-usage-ui" class="hash-link" aria-label="Direct link to Customer (End User) Usage UI" title="Direct link to Customer (End User) Usage UI">​</a></h3>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAFCAYAAAB8ZH1oAAAACXBIWXMAACxLAAAsSwGlPZapAAAAj0lEQVR4nD2OQQrCMBREc0zP4B28j5cRoe2q4Eq6cFFNq02a/3/ySip08WAYhplx8zzjvSfGSBIhpUo6EBHMFNe1LU3TsCyBEFdUDbWMqO4658x7FNw0TUdjCAFTw6wGFFPdV2IQXDUr/wnjsSTO3cB1+DAmhVIopeDqj7gmKMbNBy79i9P9udP/Vgrw1cwGJMe+tKpOUrcAAAAASUVORK5CYII=&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="334"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/customer_usage.f3faf2f.640.png" srcset="/assets/ideal-img/customer_usage.f3faf2f.640.png 640w,/assets/ideal-img/customer_usage.b1ecaef.1920.png 1920w" width="640" height="334"></noscript></div>
<p>Users can now filter usage statistics by customers, providing the same granular filtering capabilities available for teams and organizations.</p>
<p><strong>Details:</strong></p>
<ul>
<li>Filter usage analytics, spend logs, and activity metrics by customer ID</li>
<li>View customer-level breakdowns alongside existing team and user-level filters</li>
<li>Consistent filtering experience across all usage and analytics views</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-providers-and-endpoints">New Providers and Endpoints<a href="https://docs.litellm.ai/release_notes/v1-80-8#new-providers-and-endpoints" class="hash-link" aria-label="Direct link to New Providers and Endpoints" title="Direct link to New Providers and Endpoints">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="new-providers-5-new-providers">New Providers (5 new providers)<a href="https://docs.litellm.ai/release_notes/v1-80-8#new-providers-5-new-providers" class="hash-link" aria-label="Direct link to New Providers (5 new providers)" title="Direct link to New Providers (5 new providers)">​</a></h3>
<table><thead><tr><th>Provider</th><th>Supported LiteLLM Endpoints</th><th>Description</th></tr></thead><tbody><tr><td><strong><a href="https://docs.litellm.ai/docs/providers/zai">Z.AI (Zhipu AI)</a></strong></td><td><code>/v1/chat/completions</code>, <code>/v1/responses</code>, <code>/v1/messages</code></td><td>Built-in support for Zhipu AI GLM models</td></tr><tr><td><strong><a href="https://docs.litellm.ai/docs/providers/ragflow">RAGFlow</a></strong></td><td><code>/v1/chat/completions</code>, <code>/v1/responses</code>, <code>/v1/messages</code>, <code>/v1/vector_stores</code></td><td>RAG-based chat completions with vector store support</td></tr><tr><td><strong><a href="https://docs.litellm.ai/docs/providers/publicai">PublicAI</a></strong></td><td><code>/v1/chat/completions</code>, <code>/v1/responses</code>, <code>/v1/messages</code></td><td>OpenAI-compatible provider via JSON config</td></tr><tr><td><strong><a href="https://docs.litellm.ai/docs/text_to_speech">Google Cloud Chirp3 HD</a></strong></td><td><code>/v1/audio/speech</code>, <code>/v1/audio/speech/stream</code></td><td>Text-to-speech with Google Cloud Chirp3 HD voices</td></tr></tbody></table>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="new-llm-api-endpoints-2-new-endpoints">New LLM API Endpoints (2 new endpoints)<a href="https://docs.litellm.ai/release_notes/v1-80-8#new-llm-api-endpoints-2-new-endpoints" class="hash-link" aria-label="Direct link to New LLM API Endpoints (2 new endpoints)" title="Direct link to New LLM API Endpoints (2 new endpoints)">​</a></h3>
<table><thead><tr><th>Endpoint</th><th>Method</th><th>Description</th><th>Documentation</th></tr></thead><tbody><tr><td><code>/v1/agents/invoke</code></td><td>POST</td><td>Invoke A2A agents through the AI Gateway</td><td><a href="https://docs.litellm.ai/docs/a2a">Agent Gateway</a></td></tr><tr><td><code>/cursor/chat/completions</code></td><td>POST</td><td>Cursor BYOK endpoint - accepts Responses API input, returns Chat Completions output</td><td><a href="https://docs.litellm.ai/docs/tutorials/cursor_integration">Cursor Integration</a></td></tr></tbody></table>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-80-8#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support-33-new-models">New Model Support (33 new models)<a href="https://docs.litellm.ai/release_notes/v1-80-8#new-model-support-33-new-models" class="hash-link" aria-label="Direct link to New Model Support (33 new models)" title="Direct link to New Model Support (33 new models)">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th><th>Features</th></tr></thead><tbody><tr><td>OpenAI</td><td><code>gpt-5.1-codex-max</code></td><td>400K</td><td>$1.25</td><td>$10.00</td><td>Reasoning, vision, PDF input, responses API</td></tr><tr><td>Azure</td><td><code>azure/gpt-5.1-codex-max</code></td><td>400K</td><td>$1.25</td><td>$10.00</td><td>Reasoning, vision, PDF input, responses API</td></tr><tr><td>Anthropic</td><td><code>claude-opus-4-5</code></td><td>200K</td><td>$5.00</td><td>$25.00</td><td>Computer use, reasoning, vision</td></tr><tr><td>Bedrock</td><td><code>global.anthropic.claude-opus-4-5-20251101-v1:0</code></td><td>200K</td><td>$5.00</td><td>$25.00</td><td>Computer use, reasoning, vision</td></tr><tr><td>Bedrock</td><td><code>amazon.nova-2-lite-v1:0</code></td><td>1M</td><td>$0.30</td><td>$2.50</td><td>Reasoning, vision, video, PDF input</td></tr><tr><td>Bedrock</td><td><code>amazon.titan-image-generator-v2:0</code></td><td>-</td><td>-</td><td>$0.008/image</td><td>Image generation</td></tr><tr><td>Fireworks</td><td><code>fireworks_ai/deepseek-v3p2</code></td><td>164K</td><td>$1.20</td><td>$1.20</td><td>Function calling, response schema</td></tr><tr><td>Fireworks</td><td><code>fireworks_ai/kimi-k2-instruct-0905</code></td><td>262K</td><td>$0.60</td><td>$2.50</td><td>Function calling, response schema</td></tr><tr><td>DeepSeek</td><td><code>deepseek/deepseek-v3.2</code></td><td>164K</td><td>$0.28</td><td>$0.40</td><td>Reasoning, function calling</td></tr><tr><td>Mistral</td><td><code>mistral/mistral-large-3</code></td><td>256K</td><td>$0.50</td><td>$1.50</td><td>Function calling, vision</td></tr><tr><td>Azure AI</td><td><code>azure_ai/mistral-large-3</code></td><td>256K</td><td>$0.50</td><td>$1.50</td><td>Function calling, vision</td></tr><tr><td>Moonshot</td><td><code>moonshot/kimi-k2-0905-preview</code></td><td>262K</td><td>$0.60</td><td>$2.50</td><td>Function calling, web search</td></tr><tr><td>Moonshot</td><td><code>moonshot/kimi-k2-turbo-preview</code></td><td>262K</td><td>$1.15</td><td>$8.00</td><td>Function calling, web search</td></tr><tr><td>Moonshot</td><td><code>moonshot/kimi-k2-thinking-turbo</code></td><td>262K</td><td>$1.15</td><td>$8.00</td><td>Function calling, web search</td></tr><tr><td>OpenRouter</td><td><code>openrouter/deepseek/deepseek-v3.2</code></td><td>164K</td><td>$0.28</td><td>$0.40</td><td>Reasoning, function calling</td></tr><tr><td>Databricks</td><td><code>databricks/databricks-claude-haiku-4-5</code></td><td>200K</td><td>$1.00</td><td>$5.00</td><td>Reasoning, function calling</td></tr><tr><td>Databricks</td><td><code>databricks/databricks-claude-opus-4</code></td><td>200K</td><td>$15.00</td><td>$75.00</td><td>Reasoning, function calling</td></tr><tr><td>Databricks</td><td><code>databricks/databricks-claude-opus-4-1</code></td><td>200K</td><td>$15.00</td><td>$75.00</td><td>Reasoning, function calling</td></tr><tr><td>Databricks</td><td><code>databricks/databricks-claude-opus-4-5</code></td><td>200K</td><td>$5.00</td><td>$25.00</td><td>Reasoning, function calling</td></tr><tr><td>Databricks</td><td><code>databricks/databricks-claude-sonnet-4</code></td><td>200K</td><td>$3.00</td><td>$15.00</td><td>Reasoning, function calling</td></tr><tr><td>Databricks</td><td><code>databricks/databricks-claude-sonnet-4-1</code></td><td>200K</td><td>$3.00</td><td>$15.00</td><td>Reasoning, function calling</td></tr><tr><td>Databricks</td><td><code>databricks/databricks-gemini-2-5-flash</code></td><td>1M</td><td>$0.30</td><td>$2.50</td><td>Function calling</td></tr><tr><td>Databricks</td><td><code>databricks/databricks-gemini-2-5-pro</code></td><td>1M</td><td>$1.25</td><td>$10.00</td><td>Function calling</td></tr><tr><td>Databricks</td><td><code>databricks/databricks-gpt-5</code></td><td>400K</td><td>$1.25</td><td>$10.00</td><td>Function calling</td></tr><tr><td>Databricks</td><td><code>databricks/databricks-gpt-5-1</code></td><td>400K</td><td>$1.25</td><td>$10.00</td><td>Function calling</td></tr><tr><td>Databricks</td><td><code>databricks/databricks-gpt-5-mini</code></td><td>400K</td><td>$0.25</td><td>$2.00</td><td>Function calling</td></tr><tr><td>Databricks</td><td><code>databricks/databricks-gpt-5-nano</code></td><td>400K</td><td>$0.05</td><td>$0.40</td><td>Function calling</td></tr><tr><td>Vertex AI</td><td><code>vertex_ai/chirp</code></td><td>-</td><td>$30.00/1M chars</td><td>-</td><td>Text-to-speech (Chirp3 HD)</td></tr><tr><td>Z.AI</td><td><code>zai/glm-4.6</code></td><td>200K</td><td>$0.60</td><td>$2.20</td><td>Function calling</td></tr><tr><td>Z.AI</td><td><code>zai/glm-4.5</code></td><td>128K</td><td>$0.60</td><td>$2.20</td><td>Function calling</td></tr><tr><td>Z.AI</td><td><code>zai/glm-4.5v</code></td><td>128K</td><td>$0.60</td><td>$1.80</td><td>Function calling, vision</td></tr><tr><td>Z.AI</td><td><code>zai/glm-4.5-flash</code></td><td>128K</td><td>Free</td><td>Free</td><td>Function calling</td></tr><tr><td>Vertex AI</td><td><code>vertex_ai/bge-large-en-v1.5</code></td><td>-</td><td>-</td><td>-</td><td>BGE Embeddings</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-80-8#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI</a></strong></p>
<ul>
<li>Add <code>gpt-5.1-codex-max</code> model pricing and configuration - <a href="https://github.com/BerriAI/litellm/pull/17541" target="_blank" rel="noopener noreferrer">PR #17541</a></li>
<li>Add xhigh reasoning effort for gpt-5.1-codex-max - <a href="https://github.com/BerriAI/litellm/pull/17585" target="_blank" rel="noopener noreferrer">PR #17585</a></li>
<li>Add clear error message for empty LLM endpoint responses - <a href="https://github.com/BerriAI/litellm/pull/17445" target="_blank" rel="noopener noreferrer">PR #17445</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/azure/azure">Azure OpenAI</a></strong></p>
<ul>
<li>Allow reasoning_effort='none' for Azure gpt-5.1 models - <a href="https://github.com/BerriAI/litellm/pull/17311" target="_blank" rel="noopener noreferrer">PR #17311</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Add <code>claude-opus-4-5</code> alias to pricing data - <a href="https://github.com/BerriAI/litellm/pull/17313" target="_blank" rel="noopener noreferrer">PR #17313</a></li>
<li>Parse <code>&lt;budget:thinking&gt;</code> blocks for opus 4.5 - <a href="https://github.com/BerriAI/litellm/pull/17534" target="_blank" rel="noopener noreferrer">PR #17534</a></li>
<li>Update new Anthropic features as reviewed - <a href="https://github.com/BerriAI/litellm/pull/17142" target="_blank" rel="noopener noreferrer">PR #17142</a></li>
<li>Skip empty text blocks in Anthropic system messages - <a href="https://github.com/BerriAI/litellm/pull/17442" target="_blank" rel="noopener noreferrer">PR #17442</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong></p>
<ul>
<li>Add Nova embedding support - <a href="https://github.com/BerriAI/litellm/pull/17253" target="_blank" rel="noopener noreferrer">PR #17253</a></li>
<li>Add support for Bedrock Qwen 2 imported model - <a href="https://github.com/BerriAI/litellm/pull/17461" target="_blank" rel="noopener noreferrer">PR #17461</a></li>
<li>Bedrock OpenAI model support - <a href="https://github.com/BerriAI/litellm/pull/17368" target="_blank" rel="noopener noreferrer">PR #17368</a></li>
<li>Add support for file content download for Bedrock batches - <a href="https://github.com/BerriAI/litellm/pull/17470" target="_blank" rel="noopener noreferrer">PR #17470</a></li>
<li>Make streaming chunk size configurable in Bedrock API - <a href="https://github.com/BerriAI/litellm/pull/17357" target="_blank" rel="noopener noreferrer">PR #17357</a></li>
<li>Add experimental latest-user filtering for Bedrock - <a href="https://github.com/BerriAI/litellm/pull/17282" target="_blank" rel="noopener noreferrer">PR #17282</a></li>
<li>Handle Cohere v4 embed response dictionary format - <a href="https://github.com/BerriAI/litellm/pull/17220" target="_blank" rel="noopener noreferrer">PR #17220</a></li>
<li>Remove not compatible beta header from Bedrock - <a href="https://github.com/BerriAI/litellm/pull/17301" target="_blank" rel="noopener noreferrer">PR #17301</a></li>
<li>Add model price and details for Global Opus 4.5 Bedrock endpoint - <a href="https://github.com/BerriAI/litellm/pull/17380" target="_blank" rel="noopener noreferrer">PR #17380</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Gemini (Google AI Studio + Vertex AI)</a></strong></p>
<ul>
<li>Add better handling in image generation for Gemini models - <a href="https://github.com/BerriAI/litellm/pull/17292" target="_blank" rel="noopener noreferrer">PR #17292</a></li>
<li>Fix reasoning_content showing duplicate content in streaming responses - <a href="https://github.com/BerriAI/litellm/pull/17266" target="_blank" rel="noopener noreferrer">PR #17266</a></li>
<li>Handle partial JSON chunks after first valid chunk - <a href="https://github.com/BerriAI/litellm/pull/17496" target="_blank" rel="noopener noreferrer">PR #17496</a></li>
<li>Fix Gemini 3 last chunk thinking block - <a href="https://github.com/BerriAI/litellm/pull/17403" target="_blank" rel="noopener noreferrer">PR #17403</a></li>
<li>Fix Gemini image_tokens treated as text tokens in cost calculation - <a href="https://github.com/BerriAI/litellm/pull/17554" target="_blank" rel="noopener noreferrer">PR #17554</a></li>
<li>Make sure that media resolution is only for Gemini 3 model - <a href="https://github.com/BerriAI/litellm/pull/17137" target="_blank" rel="noopener noreferrer">PR #17137</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vertex">Vertex AI</a></strong></p>
<ul>
<li>Add Google Cloud Chirp3 HD support on /speech - <a href="https://github.com/BerriAI/litellm/pull/17391" target="_blank" rel="noopener noreferrer">PR #17391</a></li>
<li>Add BGE Embeddings support - <a href="https://github.com/BerriAI/litellm/pull/17362" target="_blank" rel="noopener noreferrer">PR #17362</a></li>
<li>Handle global location for Vertex AI image generation endpoint - <a href="https://github.com/BerriAI/litellm/pull/17255" target="_blank" rel="noopener noreferrer">PR #17255</a></li>
<li>Add Google Private API Endpoint to Vertex AI fields - <a href="https://github.com/BerriAI/litellm/pull/17382" target="_blank" rel="noopener noreferrer">PR #17382</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/zai">Z.AI (Zhipu AI)</a></strong></p>
<ul>
<li>Add Z.AI as built-in provider - <a href="https://github.com/BerriAI/litellm/pull/17307" target="_blank" rel="noopener noreferrer">PR #17307</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/github_copilot">GitHub Copilot</a></strong></p>
<ul>
<li>Add Embedding API support - <a href="https://github.com/BerriAI/litellm/pull/17278" target="_blank" rel="noopener noreferrer">PR #17278</a></li>
<li>Preserve encrypted_content in reasoning items for multi-turn conversations - <a href="https://github.com/BerriAI/litellm/pull/17130" target="_blank" rel="noopener noreferrer">PR #17130</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/databricks">Databricks</a></strong></p>
<ul>
<li>Update Databricks model pricing and add new models - <a href="https://github.com/BerriAI/litellm/pull/17277" target="_blank" rel="noopener noreferrer">PR #17277</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/ovhcloud">OVHcloud</a></strong></p>
<ul>
<li>Add support of audio transcription for OVHcloud - <a href="https://github.com/BerriAI/litellm/pull/17305" target="_blank" rel="noopener noreferrer">PR #17305</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/mistral">Mistral</a></strong></p>
<ul>
<li>Add Mistral Large 3 model support - <a href="https://github.com/BerriAI/litellm/pull/17547" target="_blank" rel="noopener noreferrer">PR #17547</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/moonshot">Moonshot</a></strong></p>
<ul>
<li>Fix missing Moonshot turbo models and fix incorrect pricing - <a href="https://github.com/BerriAI/litellm/pull/17432" target="_blank" rel="noopener noreferrer">PR #17432</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/togetherai">Together AI</a></strong></p>
<ul>
<li>Add context window exception mapping for Together AI - <a href="https://github.com/BerriAI/litellm/pull/17284" target="_blank" rel="noopener noreferrer">PR #17284</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/watsonx/index">WatsonX</a></strong></p>
<ul>
<li>Allow passing zen_api_key dynamically - <a href="https://github.com/BerriAI/litellm/pull/16655" target="_blank" rel="noopener noreferrer">PR #16655</a></li>
<li>Fix Watsonx Audio Transcription API - <a href="https://github.com/BerriAI/litellm/pull/17326" target="_blank" rel="noopener noreferrer">PR #17326</a></li>
<li>Fix audio transcriptions, don't force content type in request headers - <a href="https://github.com/BerriAI/litellm/pull/17546" target="_blank" rel="noopener noreferrer">PR #17546</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/fireworks_ai">Fireworks AI</a></strong></p>
<ul>
<li>Add new model <code>fireworks_ai/kimi-k2-instruct-0905</code> - <a href="https://github.com/BerriAI/litellm/pull/17328" target="_blank" rel="noopener noreferrer">PR #17328</a></li>
<li>Add <code>fireworks/deepseek-v3p2</code> - <a href="https://github.com/BerriAI/litellm/pull/17395" target="_blank" rel="noopener noreferrer">PR #17395</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/deepseek">DeepSeek</a></strong></p>
<ul>
<li>Support Deepseek 3.2 with Reasoning - <a href="https://github.com/BerriAI/litellm/pull/17384" target="_blank" rel="noopener noreferrer">PR #17384</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Nova Lite 2</a></strong></p>
<ul>
<li>Add Nova Lite 2 reasoning support with reasoningConfig - <a href="https://github.com/BerriAI/litellm/pull/17371" target="_blank" rel="noopener noreferrer">PR #17371</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/ollama">Ollama</a></strong></p>
<ul>
<li>Fix auth not working with ollama.com - <a href="https://github.com/BerriAI/litellm/pull/17191" target="_blank" rel="noopener noreferrer">PR #17191</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/groq">Groq</a></strong></p>
<ul>
<li>Fix supports_response_schema before using json_tool_call workaround - <a href="https://github.com/BerriAI/litellm/pull/17438" target="_blank" rel="noopener noreferrer">PR #17438</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vllm">vLLM</a></strong></p>
<ul>
<li>Fix empty response + vLLM streaming - <a href="https://github.com/BerriAI/litellm/pull/17516" target="_blank" rel="noopener noreferrer">PR #17516</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/azure_ai">Azure AI</a></strong></p>
<ul>
<li>Migrate Anthropic provider to Azure AI - <a href="https://github.com/BerriAI/litellm/pull/17202" target="_blank" rel="noopener noreferrer">PR #17202</a></li>
<li>Fix GA path for Azure OpenAI realtime models - <a href="https://github.com/BerriAI/litellm/pull/17260" target="_blank" rel="noopener noreferrer">PR #17260</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock#twelvelabs-pegasus---video-understanding">Bedrock TwelveLabs</a></strong></p>
<ul>
<li>Add support for TwelveLabs Pegasus video understanding - <a href="https://github.com/BerriAI/litellm/pull/17193" target="_blank" rel="noopener noreferrer">PR #17193</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-80-8#bug-fixes" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong></p>
<ul>
<li>Fix extra_headers in messages API bedrock invoke - <a href="https://github.com/BerriAI/litellm/pull/17271" target="_blank" rel="noopener noreferrer">PR #17271</a></li>
<li>Fix Bedrock models in model map - <a href="https://github.com/BerriAI/litellm/pull/17419" target="_blank" rel="noopener noreferrer">PR #17419</a></li>
<li>Make Bedrock converse messages respect modify_params as expected - <a href="https://github.com/BerriAI/litellm/pull/17427" target="_blank" rel="noopener noreferrer">PR #17427</a></li>
<li>Fix Anthropic beta headers for Bedrock imported Qwen models - <a href="https://github.com/BerriAI/litellm/pull/17467" target="_blank" rel="noopener noreferrer">PR #17467</a></li>
<li>Preserve usage from JSON response for OpenAI provider in Bedrock - <a href="https://github.com/BerriAI/litellm/pull/17589" target="_blank" rel="noopener noreferrer">PR #17589</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/sambanova">SambaNova</a></strong></p>
<ul>
<li>Fix acompletion throws error with SambaNova models - <a href="https://github.com/BerriAI/litellm/pull/17217" target="_blank" rel="noopener noreferrer">PR #17217</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Fix AttributeError when metadata is null in request body - <a href="https://github.com/BerriAI/litellm/pull/17306" target="_blank" rel="noopener noreferrer">PR #17306</a></li>
<li>Fix 500 error for malformed request - <a href="https://github.com/BerriAI/litellm/pull/17291" target="_blank" rel="noopener noreferrer">PR #17291</a></li>
<li>Respect custom LLM provider in header - <a href="https://github.com/BerriAI/litellm/pull/17290" target="_blank" rel="noopener noreferrer">PR #17290</a></li>
<li>Replace deprecated .dict() with .model_dump() in streaming_handler - <a href="https://github.com/BerriAI/litellm/pull/17359" target="_blank" rel="noopener noreferrer">PR #17359</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-80-8#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-80-8#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/response_api">Responses API</a></strong></p>
<ul>
<li>Add cost tracking for responses API - <a href="https://github.com/BerriAI/litellm/pull/17258" target="_blank" rel="noopener noreferrer">PR #17258</a></li>
<li>Map output_tokens_details of responses API to completion_tokens_details - <a href="https://github.com/BerriAI/litellm/pull/17458" target="_blank" rel="noopener noreferrer">PR #17458</a></li>
<li>Add image generation support for Responses API - <a href="https://github.com/BerriAI/litellm/pull/16586" target="_blank" rel="noopener noreferrer">PR #16586</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/batches">Batch API</a></strong></p>
<ul>
<li>Add vLLM batch+files API support - <a href="https://github.com/BerriAI/litellm/pull/15823" target="_blank" rel="noopener noreferrer">PR #15823</a></li>
<li>Fix optional parameter default value - <a href="https://github.com/BerriAI/litellm/pull/17434" target="_blank" rel="noopener noreferrer">PR #17434</a></li>
<li>Add status parameter as optional for FileObject - <a href="https://github.com/BerriAI/litellm/pull/17431" target="_blank" rel="noopener noreferrer">PR #17431</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/videos">Video Generation API</a></strong></p>
<ul>
<li>Add passthrough cost tracking for Veo - <a href="https://github.com/BerriAI/litellm/pull/17296" target="_blank" rel="noopener noreferrer">PR #17296</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/ocr">OCR API</a></strong></p>
<ul>
<li>Add missing OCR and aOCR to CallTypes enum - <a href="https://github.com/BerriAI/litellm/pull/17435" target="_blank" rel="noopener noreferrer">PR #17435</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Support routing to only websearch supported deployments - <a href="https://github.com/BerriAI/litellm/pull/17500" target="_blank" rel="noopener noreferrer">PR #17500</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs">Bugs<a href="https://docs.litellm.ai/release_notes/v1-80-8#bugs" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>General</strong>
<ul>
<li>Fix streaming error validation - <a href="https://github.com/BerriAI/litellm/pull/17242" target="_blank" rel="noopener noreferrer">PR #17242</a></li>
<li>Add length validation for empty tool_calls in delta - <a href="https://github.com/BerriAI/litellm/pull/17523" target="_blank" rel="noopener noreferrer">PR #17523</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-80-8#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-80-8#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong>New Login Page</strong></p>
<ul>
<li>New Login Page UI - <a href="https://github.com/BerriAI/litellm/pull/17443" target="_blank" rel="noopener noreferrer">PR #17443</a></li>
<li>Refactor /login route - <a href="https://github.com/BerriAI/litellm/pull/17379" target="_blank" rel="noopener noreferrer">PR #17379</a></li>
<li>Add auto_redirect_to_sso to UI Config - <a href="https://github.com/BerriAI/litellm/pull/17399" target="_blank" rel="noopener noreferrer">PR #17399</a></li>
<li>Add Auto Redirect to SSO to New Login Page - <a href="https://github.com/BerriAI/litellm/pull/17451" target="_blank" rel="noopener noreferrer">PR #17451</a></li>
</ul>
</li>
<li>
<p><strong>Customer (End User) Usage</strong></p>
<ul>
<li>Customer (end user) Usage feature - <a href="https://github.com/BerriAI/litellm/pull/17498" target="_blank" rel="noopener noreferrer">PR #17498</a></li>
<li>Customer Usage UI - <a href="https://github.com/BerriAI/litellm/pull/17506" target="_blank" rel="noopener noreferrer">PR #17506</a></li>
<li>Add Info Banner for Customer Usage - <a href="https://github.com/BerriAI/litellm/pull/17598" target="_blank" rel="noopener noreferrer">PR #17598</a></li>
</ul>
</li>
<li>
<p><strong>Virtual Keys</strong></p>
<ul>
<li>Standardize API Key vs Virtual Key in UI - <a href="https://github.com/BerriAI/litellm/pull/17325" target="_blank" rel="noopener noreferrer">PR #17325</a></li>
<li>Add User Alias Column to Internal User Table - <a href="https://github.com/BerriAI/litellm/pull/17321" target="_blank" rel="noopener noreferrer">PR #17321</a></li>
<li>Delete Credential Enhancements - <a href="https://github.com/BerriAI/litellm/pull/17317" target="_blank" rel="noopener noreferrer">PR #17317</a></li>
</ul>
</li>
<li>
<p><strong>Models + Endpoints</strong></p>
<ul>
<li>Show all credential values on Edit Credential Modal - <a href="https://github.com/BerriAI/litellm/pull/17397" target="_blank" rel="noopener noreferrer">PR #17397</a></li>
<li>Change Edit Team Models Shown to Match Create Team - <a href="https://github.com/BerriAI/litellm/pull/17394" target="_blank" rel="noopener noreferrer">PR #17394</a></li>
<li>Support Images in Compare UI - <a href="https://github.com/BerriAI/litellm/pull/17562" target="_blank" rel="noopener noreferrer">PR #17562</a></li>
</ul>
</li>
<li>
<p><strong>Callbacks</strong></p>
<ul>
<li>Show all callbacks on UI - <a href="https://github.com/BerriAI/litellm/pull/16335" target="_blank" rel="noopener noreferrer">PR #16335</a></li>
<li>Credentials to use React Query - <a href="https://github.com/BerriAI/litellm/pull/17465" target="_blank" rel="noopener noreferrer">PR #17465</a></li>
</ul>
</li>
<li>
<p><strong>Management Routes</strong></p>
<ul>
<li>Allow admin viewer to access global tag usage - <a href="https://github.com/BerriAI/litellm/pull/17501" target="_blank" rel="noopener noreferrer">PR #17501</a></li>
<li>Allow wildcard routes for nonproxy admin (SCIM) - <a href="https://github.com/BerriAI/litellm/pull/17178" target="_blank" rel="noopener noreferrer">PR #17178</a></li>
<li>Return 404 when a user is not found on /user/info - <a href="https://github.com/BerriAI/litellm/pull/16850" target="_blank" rel="noopener noreferrer">PR #16850</a></li>
</ul>
</li>
<li>
<p><strong>OCI Configuration</strong></p>
<ul>
<li>Enable Oracle Cloud Infrastructure configuration via UI - <a href="https://github.com/BerriAI/litellm/pull/17159" target="_blank" rel="noopener noreferrer">PR #17159</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs-1">Bugs<a href="https://docs.litellm.ai/release_notes/v1-80-8#bugs-1" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li>
<p><strong>UI Fixes</strong></p>
<ul>
<li>Fix Request and Response Panel JSONViewer - <a href="https://github.com/BerriAI/litellm/pull/17233" target="_blank" rel="noopener noreferrer">PR #17233</a></li>
<li>Adding Button Loading States to Edit Settings - <a href="https://github.com/BerriAI/litellm/pull/17236" target="_blank" rel="noopener noreferrer">PR #17236</a></li>
<li>Fix Various Text, button state, and test changes - <a href="https://github.com/BerriAI/litellm/pull/17237" target="_blank" rel="noopener noreferrer">PR #17237</a></li>
<li>Fix Fallbacks Immediately Deleting before API resolves - <a href="https://github.com/BerriAI/litellm/pull/17238" target="_blank" rel="noopener noreferrer">PR #17238</a></li>
<li>Remove Feature Flags - <a href="https://github.com/BerriAI/litellm/pull/17240" target="_blank" rel="noopener noreferrer">PR #17240</a></li>
<li>Fix metadata tags and model name display in UI for Azure passthrough - <a href="https://github.com/BerriAI/litellm/pull/17258" target="_blank" rel="noopener noreferrer">PR #17258</a></li>
<li>Change labeling around Vertex Fields - <a href="https://github.com/BerriAI/litellm/pull/17383" target="_blank" rel="noopener noreferrer">PR #17383</a></li>
<li>Remove second scrollbar when sidebar is expanded + tooltip z index - <a href="https://github.com/BerriAI/litellm/pull/17436" target="_blank" rel="noopener noreferrer">PR #17436</a></li>
<li>Fix Select in Edit Membership Modal - <a href="https://github.com/BerriAI/litellm/pull/17524" target="_blank" rel="noopener noreferrer">PR #17524</a></li>
<li>Change useAuthorized Hook to redirect to new Login Page - <a href="https://github.com/BerriAI/litellm/pull/17553" target="_blank" rel="noopener noreferrer">PR #17553</a></li>
</ul>
</li>
<li>
<p><strong>SSO</strong></p>
<ul>
<li>Fix the generic SSO provider - <a href="https://github.com/BerriAI/litellm/pull/17227" target="_blank" rel="noopener noreferrer">PR #17227</a></li>
<li>Clear SSO integration for all users - <a href="https://github.com/BerriAI/litellm/pull/17287" target="_blank" rel="noopener noreferrer">PR #17287</a></li>
<li>Fix SSO users not added to Entra synced team - <a href="https://github.com/BerriAI/litellm/pull/17331" target="_blank" rel="noopener noreferrer">PR #17331</a></li>
</ul>
</li>
<li>
<p><strong>Auth / JWT</strong></p>
<ul>
<li>JWT Auth - Allow using regular OIDC flow with user info endpoints - <a href="https://github.com/BerriAI/litellm/pull/17324" target="_blank" rel="noopener noreferrer">PR #17324</a></li>
<li>Fix litellm user auth not passing issue - <a href="https://github.com/BerriAI/litellm/pull/17342" target="_blank" rel="noopener noreferrer">PR #17342</a></li>
<li>Add other routes in JWT auth - <a href="https://github.com/BerriAI/litellm/pull/17345" target="_blank" rel="noopener noreferrer">PR #17345</a></li>
<li>Fix new org team validate against org - <a href="https://github.com/BerriAI/litellm/pull/17333" target="_blank" rel="noopener noreferrer">PR #17333</a></li>
<li>Fix litellm_enterprise ensure imported routes exist - <a href="https://github.com/BerriAI/litellm/pull/17337" target="_blank" rel="noopener noreferrer">PR #17337</a></li>
<li>Use organization.members instead of deprecated organization field - <a href="https://github.com/BerriAI/litellm/pull/17557" target="_blank" rel="noopener noreferrer">PR #17557</a></li>
</ul>
</li>
<li>
<p><strong>Organizations/Teams</strong></p>
<ul>
<li>Fix organization max budget not enforced - <a href="https://github.com/BerriAI/litellm/pull/17334" target="_blank" rel="noopener noreferrer">PR #17334</a></li>
<li>Fix budget update to allow null max_budget - <a href="https://github.com/BerriAI/litellm/pull/17545" target="_blank" rel="noopener noreferrer">PR #17545</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="ai-integrations-2-new-integrations">AI Integrations (2 new integrations)<a href="https://docs.litellm.ai/release_notes/v1-80-8#ai-integrations-2-new-integrations" class="hash-link" aria-label="Direct link to AI Integrations (2 new integrations)" title="Direct link to AI Integrations (2 new integrations)">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="logging-1-new-integration">Logging (1 new integration)<a href="https://docs.litellm.ai/release_notes/v1-80-8#logging-1-new-integration" class="hash-link" aria-label="Direct link to Logging (1 new integration)" title="Direct link to Logging (1 new integration)">​</a></h3>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-integration">New Integration<a href="https://docs.litellm.ai/release_notes/v1-80-8#new-integration" class="hash-link" aria-label="Direct link to New Integration" title="Direct link to New Integration">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging">Weave</a></strong>
<ul>
<li>Basic Weave OTEL integration - <a href="https://github.com/BerriAI/litellm/pull/17439" target="_blank" rel="noopener noreferrer">PR #17439</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="improvements--fixes">Improvements &amp; Fixes<a href="https://docs.litellm.ai/release_notes/v1-80-8#improvements--fixes" class="hash-link" aria-label="Direct link to Improvements &amp; Fixes" title="Direct link to Improvements &amp; Fixes">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#datadog">DataDog</a></strong></p>
<ul>
<li>Fix Datadog callback regression when ddtrace is installed - <a href="https://github.com/BerriAI/litellm/pull/17393" target="_blank" rel="noopener noreferrer">PR #17393</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/observability/arize_integration">Arize Phoenix</a></strong></p>
<ul>
<li>Fix clean arize-phoenix traces - <a href="https://github.com/BerriAI/litellm/pull/16611" target="_blank" rel="noopener noreferrer">PR #16611</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#mlflow">MLflow</a></strong></p>
<ul>
<li>Fix MLflow streaming spans for Anthropic passthrough - <a href="https://github.com/BerriAI/litellm/pull/17288" target="_blank" rel="noopener noreferrer">PR #17288</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#langfuse">Langfuse</a></strong></p>
<ul>
<li>Fix Langfuse logger test mock setup - <a href="https://github.com/BerriAI/litellm/pull/17591" target="_blank" rel="noopener noreferrer">PR #17591</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Improve PII anonymization handling in logging callbacks - <a href="https://github.com/BerriAI/litellm/pull/17207" target="_blank" rel="noopener noreferrer">PR #17207</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails-1-new-integration">Guardrails (1 new integration)<a href="https://docs.litellm.ai/release_notes/v1-80-8#guardrails-1-new-integration" class="hash-link" aria-label="Direct link to Guardrails (1 new integration)" title="Direct link to Guardrails (1 new integration)">​</a></h3>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-integration-1">New Integration<a href="https://docs.litellm.ai/release_notes/v1-80-8#new-integration-1" class="hash-link" aria-label="Direct link to New Integration" title="Direct link to New Integration">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/adding_provider/generic_guardrail_api">Generic Guardrail API</a></strong>
<ul>
<li>Generic Guardrail API - allows guardrail providers to add INSTANT support for LiteLLM w/out PR to repo - <a href="https://github.com/BerriAI/litellm/pull/17175" target="_blank" rel="noopener noreferrer">PR #17175</a></li>
<li>Guardrails API V2 - user api key metadata, session id, specify input type (request/response), image support - <a href="https://github.com/BerriAI/litellm/pull/17338" target="_blank" rel="noopener noreferrer">PR #17338</a></li>
<li>Guardrails API - add streaming support - <a href="https://github.com/BerriAI/litellm/pull/17400" target="_blank" rel="noopener noreferrer">PR #17400</a></li>
<li>Guardrails API - support tool call checks on OpenAI <code>/chat/completions</code>, OpenAI <code>/responses</code>, Anthropic <code>/v1/messages</code> - <a href="https://github.com/BerriAI/litellm/pull/17459" target="_blank" rel="noopener noreferrer">PR #17459</a></li>
<li>Guardrails API - new <code>structured_messages</code> param - <a href="https://github.com/BerriAI/litellm/pull/17518" target="_blank" rel="noopener noreferrer">PR #17518</a></li>
<li>Correctly map a v1/messages call to the anthropic unified guardrail - <a href="https://github.com/BerriAI/litellm/pull/17424" target="_blank" rel="noopener noreferrer">PR #17424</a></li>
<li>Support during_call event type for unified guardrails - <a href="https://github.com/BerriAI/litellm/pull/17514" target="_blank" rel="noopener noreferrer">PR #17514</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="improvements--fixes-1">Improvements &amp; Fixes<a href="https://docs.litellm.ai/release_notes/v1-80-8#improvements--fixes-1" class="hash-link" aria-label="Direct link to Improvements &amp; Fixes" title="Direct link to Improvements &amp; Fixes">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/noma_security">Noma Guardrail</a></strong></p>
<ul>
<li>Refactor Noma guardrail to use shared Responses transformation and include system instructions - <a href="https://github.com/BerriAI/litellm/pull/17315" target="_blank" rel="noopener noreferrer">PR #17315</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/pii_masking_v2">Presidio</a></strong></p>
<ul>
<li>Handle empty content and error dict responses in guardrails - <a href="https://github.com/BerriAI/litellm/pull/17489" target="_blank" rel="noopener noreferrer">PR #17489</a></li>
<li>Fix Presidio guardrail test TypeError and license base64 decoding error - <a href="https://github.com/BerriAI/litellm/pull/17538" target="_blank" rel="noopener noreferrer">PR #17538</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/tool_permission">Tool Permissions</a></strong></p>
<ul>
<li>Add regex-based tool_name/tool_type matching for tool-permission - <a href="https://github.com/BerriAI/litellm/pull/17164" target="_blank" rel="noopener noreferrer">PR #17164</a></li>
<li>Add images for tool permission guardrail documentation - <a href="https://github.com/BerriAI/litellm/pull/17322" target="_blank" rel="noopener noreferrer">PR #17322</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/aim_security">AIM Guardrails</a></strong></p>
<ul>
<li>Fix AIM guardrail tests - <a href="https://github.com/BerriAI/litellm/pull/17499" target="_blank" rel="noopener noreferrer">PR #17499</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails/bedrock">Bedrock Guardrails</a></strong></p>
<ul>
<li>Fix Bedrock Guardrail indent and import - <a href="https://github.com/BerriAI/litellm/pull/17378" target="_blank" rel="noopener noreferrer">PR #17378</a></li>
</ul>
</li>
<li>
<p><strong>General Guardrails</strong></p>
<ul>
<li>Mask all matching keywords in content filter - <a href="https://github.com/BerriAI/litellm/pull/17521" target="_blank" rel="noopener noreferrer">PR #17521</a></li>
<li>Ensure guardrail metadata is preserved in request_data - <a href="https://github.com/BerriAI/litellm/pull/17593" target="_blank" rel="noopener noreferrer">PR #17593</a></li>
<li>Fix apply_guardrail method and improve test isolation - <a href="https://github.com/BerriAI/litellm/pull/17555" target="_blank" rel="noopener noreferrer">PR #17555</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="secret-managers">Secret Managers<a href="https://docs.litellm.ai/release_notes/v1-80-8#secret-managers" class="hash-link" aria-label="Direct link to Secret Managers" title="Direct link to Secret Managers">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/secret_managers/cyberark">CyberArk</a></strong></p>
<ul>
<li>Allow setting SSL verify to false - <a href="https://github.com/BerriAI/litellm/pull/17433" target="_blank" rel="noopener noreferrer">PR #17433</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Make email and secret manager operations independent in key management hooks - <a href="https://github.com/BerriAI/litellm/pull/17551" target="_blank" rel="noopener noreferrer">PR #17551</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-80-8#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li>
<p><strong>Rate Limiting</strong></p>
<ul>
<li>Parallel Request Limiter with /messages - <a href="https://github.com/BerriAI/litellm/pull/17426" target="_blank" rel="noopener noreferrer">PR #17426</a></li>
<li>Allow using dynamic rate limit/priority reservation on teams - <a href="https://github.com/BerriAI/litellm/pull/17061" target="_blank" rel="noopener noreferrer">PR #17061</a></li>
<li>Dynamic Rate Limiter - Fix token count increases/decreases by 1 instead of actual count + Redis TTL - <a href="https://github.com/BerriAI/litellm/pull/17558" target="_blank" rel="noopener noreferrer">PR #17558</a></li>
</ul>
</li>
<li>
<p><strong>Spend Logs</strong></p>
<ul>
<li>Deprecate <code>spend/logs</code> &amp; add <code>spend/logs/v2</code> - <a href="https://github.com/BerriAI/litellm/pull/17167" target="_blank" rel="noopener noreferrer">PR #17167</a></li>
<li>Optimize SpendLogs queries to use timestamp filtering for index usage - <a href="https://github.com/BerriAI/litellm/pull/17504" target="_blank" rel="noopener noreferrer">PR #17504</a></li>
</ul>
</li>
<li>
<p><strong>Enforce User Param</strong></p>
<ul>
<li>Enforce support of enforce_user_param to OpenAI post endpoints - <a href="https://github.com/BerriAI/litellm/pull/17407" target="_blank" rel="noopener noreferrer">PR #17407</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-gateway">MCP Gateway<a href="https://docs.litellm.ai/release_notes/v1-80-8#mcp-gateway" class="hash-link" aria-label="Direct link to MCP Gateway" title="Direct link to MCP Gateway">​</a></h2>
<ul>
<li>
<p><strong>MCP Configuration</strong></p>
<ul>
<li>Remove URL format validation for MCP server endpoints - <a href="https://github.com/BerriAI/litellm/pull/17270" target="_blank" rel="noopener noreferrer">PR #17270</a></li>
<li>Add stack trace to MCP error message - <a href="https://github.com/BerriAI/litellm/pull/17269" target="_blank" rel="noopener noreferrer">PR #17269</a></li>
</ul>
</li>
<li>
<p><strong>MCP Tool Results</strong></p>
<ul>
<li>Preserve tool metadata in CallToolResult - <a href="https://github.com/BerriAI/litellm/pull/17561" target="_blank" rel="noopener noreferrer">PR #17561</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="agent-gateway-a2a-1">Agent Gateway (A2A)<a href="https://docs.litellm.ai/release_notes/v1-80-8#agent-gateway-a2a-1" class="hash-link" aria-label="Direct link to Agent Gateway (A2A)" title="Direct link to Agent Gateway (A2A)">​</a></h2>
<ul>
<li>
<p><strong>Agent Invocation</strong></p>
<ul>
<li>Allow invoking agents through AI Gateway - <a href="https://github.com/BerriAI/litellm/pull/17440" target="_blank" rel="noopener noreferrer">PR #17440</a></li>
<li>Allow tracking request/response in "Logs" Page - <a href="https://github.com/BerriAI/litellm/pull/17449" target="_blank" rel="noopener noreferrer">PR #17449</a></li>
</ul>
</li>
<li>
<p><strong>Agent Access Control</strong></p>
<ul>
<li>Enforce Allowed agents by key, team + add agent access groups on backend - <a href="https://github.com/BerriAI/litellm/pull/17502" target="_blank" rel="noopener noreferrer">PR #17502</a></li>
</ul>
</li>
<li>
<p><strong>Agent Gateway UI</strong></p>
<ul>
<li>Allow testing agents on UI - <a href="https://github.com/BerriAI/litellm/pull/17455" target="_blank" rel="noopener noreferrer">PR #17455</a></li>
<li>Set allowed agents by key, team - <a href="https://github.com/BerriAI/litellm/pull/17511" target="_blank" rel="noopener noreferrer">PR #17511</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements">Performance / Loadbalancing / Reliability improvements<a href="https://docs.litellm.ai/release_notes/v1-80-8#performance--loadbalancing--reliability-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements" title="Direct link to Performance / Loadbalancing / Reliability improvements">​</a></h2>
<ul>
<li>
<p><strong>Audio/Speech Performance</strong></p>
<ul>
<li>Fix <code>/audio/speech</code> performance by using <code>shared_sessions</code> - <a href="https://github.com/BerriAI/litellm/pull/16739" target="_blank" rel="noopener noreferrer">PR #16739</a></li>
</ul>
</li>
<li>
<p><strong>Memory Optimization</strong></p>
<ul>
<li>Prevent memory leak in aiohttp connection pooling - <a href="https://github.com/BerriAI/litellm/pull/17388" target="_blank" rel="noopener noreferrer">PR #17388</a></li>
<li>Lazy-load utils to reduce memory + import time - <a href="https://github.com/BerriAI/litellm/pull/17171" target="_blank" rel="noopener noreferrer">PR #17171</a></li>
</ul>
</li>
<li>
<p><strong>Database</strong></p>
<ul>
<li>Update default database connection number - <a href="https://github.com/BerriAI/litellm/pull/17353" target="_blank" rel="noopener noreferrer">PR #17353</a></li>
<li>Update default proxy_batch_write_at number - <a href="https://github.com/BerriAI/litellm/pull/17355" target="_blank" rel="noopener noreferrer">PR #17355</a></li>
<li>Add background health checks to db - <a href="https://github.com/BerriAI/litellm/pull/17528" target="_blank" rel="noopener noreferrer">PR #17528</a></li>
</ul>
</li>
<li>
<p><strong>Proxy Caching</strong></p>
<ul>
<li>Fix proxy caching between requests in aiohttp transport - <a href="https://github.com/BerriAI/litellm/pull/17122" target="_blank" rel="noopener noreferrer">PR #17122</a></li>
</ul>
</li>
<li>
<p><strong>Session Management</strong></p>
<ul>
<li>Fix session consistency, move Lasso API version away from source code - <a href="https://github.com/BerriAI/litellm/pull/17316" target="_blank" rel="noopener noreferrer">PR #17316</a></li>
<li>Conditionally pass enable_cleanup_closed to aiohttp TCPConnector - <a href="https://github.com/BerriAI/litellm/pull/17367" target="_blank" rel="noopener noreferrer">PR #17367</a></li>
</ul>
</li>
<li>
<p><strong>Vector Store</strong></p>
<ul>
<li>Fix vector store configuration synchronization failure - <a href="https://github.com/BerriAI/litellm/pull/17525" target="_blank" rel="noopener noreferrer">PR #17525</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates">Documentation Updates<a href="https://docs.litellm.ai/release_notes/v1-80-8#documentation-updates" class="hash-link" aria-label="Direct link to Documentation Updates" title="Direct link to Documentation Updates">​</a></h2>
<ul>
<li>
<p><strong>Provider Documentation</strong></p>
<ul>
<li>Add Azure AI Foundry documentation for Claude models - <a href="https://github.com/BerriAI/litellm/pull/17104" target="_blank" rel="noopener noreferrer">PR #17104</a></li>
<li>Document responses and embedding API for GitHub Copilot - <a href="https://github.com/BerriAI/litellm/pull/17456" target="_blank" rel="noopener noreferrer">PR #17456</a></li>
<li>Add gpt-5.1-codex-max to OpenAI provider documentation - <a href="https://github.com/BerriAI/litellm/pull/17602" target="_blank" rel="noopener noreferrer">PR #17602</a></li>
<li>Update Instructions For Phoenix Integration - <a href="https://github.com/BerriAI/litellm/pull/17373" target="_blank" rel="noopener noreferrer">PR #17373</a></li>
</ul>
</li>
<li>
<p><strong>Guides</strong></p>
<ul>
<li>Add guide on how to debug gateway error vs provider error - <a href="https://github.com/BerriAI/litellm/pull/17387" target="_blank" rel="noopener noreferrer">PR #17387</a></li>
<li>Agent Gateway documentation - <a href="https://github.com/BerriAI/litellm/pull/17454" target="_blank" rel="noopener noreferrer">PR #17454</a></li>
<li>A2A Permission management documentation - <a href="https://github.com/BerriAI/litellm/pull/17515" target="_blank" rel="noopener noreferrer">PR #17515</a></li>
<li>Update docs to link agent hub - <a href="https://github.com/BerriAI/litellm/pull/17462" target="_blank" rel="noopener noreferrer">PR #17462</a></li>
</ul>
</li>
<li>
<p><strong>Projects</strong></p>
<ul>
<li>Add Google ADK and Harbor to projects - <a href="https://github.com/BerriAI/litellm/pull/17352" target="_blank" rel="noopener noreferrer">PR #17352</a></li>
<li>Add Microsoft Agent Lightning to projects - <a href="https://github.com/BerriAI/litellm/pull/17422" target="_blank" rel="noopener noreferrer">PR #17422</a></li>
</ul>
</li>
<li>
<p><strong>Cleanup</strong></p>
<ul>
<li>Cleanup: Remove orphan docs pages and Docusaurus template files - <a href="https://github.com/BerriAI/litellm/pull/17356" target="_blank" rel="noopener noreferrer">PR #17356</a></li>
<li>Remove <code>source .env</code> from docs - <a href="https://github.com/BerriAI/litellm/pull/17466" target="_blank" rel="noopener noreferrer">PR #17466</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="infrastructure--cicd">Infrastructure / CI/CD<a href="https://docs.litellm.ai/release_notes/v1-80-8#infrastructure--cicd" class="hash-link" aria-label="Direct link to Infrastructure / CI/CD" title="Direct link to Infrastructure / CI/CD">​</a></h2>
<ul>
<li>
<p><strong>Helm Chart</strong></p>
<ul>
<li>Add ingress-only labels - <a href="https://github.com/BerriAI/litellm/pull/17348" target="_blank" rel="noopener noreferrer">PR #17348</a></li>
</ul>
</li>
<li>
<p><strong>Docker</strong></p>
<ul>
<li>Add retry logic to apk package installation in Dockerfile.non_root - <a href="https://github.com/BerriAI/litellm/pull/17596" target="_blank" rel="noopener noreferrer">PR #17596</a></li>
<li>Chainguard fixes - <a href="https://github.com/BerriAI/litellm/pull/17406" target="_blank" rel="noopener noreferrer">PR #17406</a></li>
</ul>
</li>
<li>
<p><strong>OpenAPI Schema</strong></p>
<ul>
<li>Refactor add_schema_to_components to move definitions to components/schemas - <a href="https://github.com/BerriAI/litellm/pull/17389" target="_blank" rel="noopener noreferrer">PR #17389</a></li>
</ul>
</li>
<li>
<p><strong>Security</strong></p>
<ul>
<li>Fix security vulnerability: update mdast-util-to-hast to 13.2.1 - <a href="https://github.com/BerriAI/litellm/pull/17601" target="_blank" rel="noopener noreferrer">PR #17601</a></li>
<li>Bump jws from 3.2.2 to 3.2.3 - <a href="https://github.com/BerriAI/litellm/pull/17494" target="_blank" rel="noopener noreferrer">PR #17494</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-80-8#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@weichiet made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17242" target="_blank" rel="noopener noreferrer">PR #17242</a></li>
<li>@AndyForest made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17220" target="_blank" rel="noopener noreferrer">PR #17220</a></li>
<li>@omkar806 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17217" target="_blank" rel="noopener noreferrer">PR #17217</a></li>
<li>@v0rtex20k made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17178" target="_blank" rel="noopener noreferrer">PR #17178</a></li>
<li>@hxomer made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17207" target="_blank" rel="noopener noreferrer">PR #17207</a></li>
<li>@orgersh92 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17316" target="_blank" rel="noopener noreferrer">PR #17316</a></li>
<li>@dannykopping made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17313" target="_blank" rel="noopener noreferrer">PR #17313</a></li>
<li>@rioiart made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17333" target="_blank" rel="noopener noreferrer">PR #17333</a></li>
<li>@codgician made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17278" target="_blank" rel="noopener noreferrer">PR #17278</a></li>
<li>@epistoteles made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17277" target="_blank" rel="noopener noreferrer">PR #17277</a></li>
<li>@kothamah made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17368" target="_blank" rel="noopener noreferrer">PR #17368</a></li>
<li>@flozonn made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17371" target="_blank" rel="noopener noreferrer">PR #17371</a></li>
<li>@richardmcsong made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17389" target="_blank" rel="noopener noreferrer">PR #17389</a></li>
<li>@matt-greathouse made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17384" target="_blank" rel="noopener noreferrer">PR #17384</a></li>
<li>@mossbanay made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17380" target="_blank" rel="noopener noreferrer">PR #17380</a></li>
<li>@mhielpos-asapp made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17376" target="_blank" rel="noopener noreferrer">PR #17376</a></li>
<li>@Joilence made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17367" target="_blank" rel="noopener noreferrer">PR #17367</a></li>
<li>@deepaktammali made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17357" target="_blank" rel="noopener noreferrer">PR #17357</a></li>
<li>@axiomofjoy made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16611" target="_blank" rel="noopener noreferrer">PR #16611</a></li>
<li>@DevajMody made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17445" target="_blank" rel="noopener noreferrer">PR #17445</a></li>
<li>@andrewtruong made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17439" target="_blank" rel="noopener noreferrer">PR #17439</a></li>
<li>@AnasAbdelR made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17490" target="_blank" rel="noopener noreferrer">PR #17490</a></li>
<li>@dominicfeliton made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17516" target="_blank" rel="noopener noreferrer">PR #17516</a></li>
<li>@kristianmitk made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17504" target="_blank" rel="noopener noreferrer">PR #17504</a></li>
<li>@rgshr made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17130" target="_blank" rel="noopener noreferrer">PR #17130</a></li>
<li>@dominicfallows made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17489" target="_blank" rel="noopener noreferrer">PR #17489</a></li>
<li>@irfansofyana made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17467" target="_blank" rel="noopener noreferrer">PR #17467</a></li>
<li>@GusBricker made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17191" target="_blank" rel="noopener noreferrer">PR #17191</a></li>
<li>@OlivverX made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17255" target="_blank" rel="noopener noreferrer">PR #17255</a></li>
<li>@withsmilo made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/17585" target="_blank" rel="noopener noreferrer">PR #17585</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog">Full Changelog<a href="https://docs.litellm.ai/release_notes/v1-80-8#full-changelog" class="hash-link" aria-label="Direct link to Full Changelog" title="Direct link to Full Changelog">​</a></h2>
<p><strong><a href="https://github.com/BerriAI/litellm/compare/v1.80.7-nightly...v1.80.8" target="_blank" rel="noopener noreferrer">View complete changelog on GitHub</a></strong></p>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[v1.80.5-stable - Gemini 3.0 Support]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-80-5</id>
        <link href="https://docs.litellm.ai/release_notes/v1-80-5"/>
        <updated>2025-11-22T10:00:00.000Z</updated>
        <summary type="html"><![CDATA[Deploy this version]]></summary>
        <content type="html"><![CDATA[<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-80-5#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">docker.litellm.ai/berriai/litellm:v1.80.5-stable</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.80.5</span><br></span></code></pre></div></div></div></div></div>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-80-5#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<ul>
<li><strong>Gemini 3</strong> - <a href="https://docs.litellm.ai/blog/gemini_3">Day-0 support for Gemini 3 models with thought signatures</a></li>
<li><strong>Prompt Management</strong> - <a href="https://docs.litellm.ai/docs/proxy/litellm_prompt_management">Full prompt versioning support with UI for editing, testing, and version history</a></li>
<li><strong>MCP Hub</strong> - <a href="https://docs.litellm.ai/docs/proxy/ai_hub#mcp-servers">Publish and discover MCP servers within your organization</a></li>
<li><strong>Model Compare UI</strong> - <a href="https://docs.litellm.ai/docs/proxy/model_compare_ui">Side-by-side model comparison interface for testing</a></li>
<li><strong>Batch API Spend Tracking</strong> - <a href="https://docs.litellm.ai/docs/proxy/cost_tracking#-custom-spend-log-metadata">Granular spend tracking with custom metadata for batch and file creation requests</a></li>
<li><strong>AWS IAM Secret Manager</strong> - <a href="https://docs.litellm.ai/docs/secret_managers/aws_secret_manager#iam-role-assumption">IAM role authentication support for AWS Secret Manager</a></li>
<li><strong>Logging Callback Controls</strong> - <a href="https://docs.litellm.ai/docs/proxy/dynamic_logging#disabling-dynamic-callback-management-enterprise">Admin-level controls to prevent callers from disabling logging callbacks in compliance environments</a></li>
<li><strong>Proxy CLI JWT Authentication</strong> - <a href="https://docs.litellm.ai/docs/proxy/cli_sso">Enable developers to authenticate to LiteLLM AI Gateway using the Proxy CLI</a></li>
<li><strong>Batch API Routing</strong> - <a href="https://docs.litellm.ai/docs/batches#multi-account--model-based-routing">Route batch operations to different provider accounts using model-specific credentials from your config.yaml</a></li>
</ul>
<hr>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="prompt-management">Prompt Management<a href="https://docs.litellm.ai/release_notes/v1-80-5#prompt-management" class="hash-link" aria-label="Direct link to Prompt Management" title="Direct link to Prompt Management">​</a></h3>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAFCAYAAAB8ZH1oAAAACXBIWXMAACxLAAAsSwGlPZapAAAAfElEQVR4nEWNywqEMBRD/f9PdONK0S66kLG9zzlDFZlAIIQ8pqMU1nVl33dKOei94+6YGWaKqt562raNeZ5ZloVaKyKCe+ARqBqiSmYyjcbguyKiRCSizvnpXF2I/DI9Fw/H7Si9MA/qKVyW/+BYbK1ztYa5ExG3H+5kJD9A2cJGDm7IFwAAAABJRU5ErkJggg==&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="334"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/prompt_history.954be8c.640.png" srcset="/assets/ideal-img/prompt_history.954be8c.640.png 640w,/assets/ideal-img/prompt_history.27a77e7.1920.png 1920w" width="640" height="334"></noscript></div>
<br>
<br>
<p>This release introduces <strong>LiteLLM Prompt Studio</strong> - a comprehensive prompt management solution built directly into the LiteLLM UI. Create, test, and version your prompts without leaving your browser.</p>
<p>You can now do the following on LiteLLM Prompt Studio:</p>
<ul>
<li><strong>Create &amp; Test Prompts</strong>: Build prompts with developer messages (system instructions) and test them in real-time with an interactive chat interface</li>
<li><strong>Dynamic Variables</strong>: Use <code>{{variable_name}}</code> syntax to create reusable prompt templates with automatic variable detection</li>
<li><strong>Version Control</strong>: Automatic versioning for every prompt update with complete version history tracking and rollback capabilities</li>
<li><strong>Prompt Studio</strong>: Edit prompts in a dedicated studio environment with live testing and preview</li>
</ul>
<p><strong>API Integration:</strong></p>
<p>Use your prompts in any application with simple API calls:</p>
<div class="language-python codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-python codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">response </span><span class="token operator" style="color:#393A34">=</span><span class="token plain"> client</span><span class="token punctuation" style="color:#393A34">.</span><span class="token plain">chat</span><span class="token punctuation" style="color:#393A34">.</span><span class="token plain">completions</span><span class="token punctuation" style="color:#393A34">.</span><span class="token plain">create</span><span class="token punctuation" style="color:#393A34">(</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">    model</span><span class="token operator" style="color:#393A34">=</span><span class="token string" style="color:#e3116c">"gpt-4"</span><span class="token punctuation" style="color:#393A34">,</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">    extra_body</span><span class="token operator" style="color:#393A34">=</span><span class="token punctuation" style="color:#393A34">{</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">        </span><span class="token string" style="color:#e3116c">"prompt_id"</span><span class="token punctuation" style="color:#393A34">:</span><span class="token plain"> </span><span class="token string" style="color:#e3116c">"your-prompt-id"</span><span class="token punctuation" style="color:#393A34">,</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">        </span><span class="token string" style="color:#e3116c">"prompt_version"</span><span class="token punctuation" style="color:#393A34">:</span><span class="token plain"> </span><span class="token number" style="color:#36acaa">2</span><span class="token punctuation" style="color:#393A34">,</span><span class="token plain">  </span><span class="token comment" style="color:#999988;font-style:italic"># Optional: specify version</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">        </span><span class="token string" style="color:#e3116c">"prompt_variables"</span><span class="token punctuation" style="color:#393A34">:</span><span class="token plain"> </span><span class="token punctuation" style="color:#393A34">{</span><span class="token string" style="color:#e3116c">"name"</span><span class="token punctuation" style="color:#393A34">:</span><span class="token plain"> </span><span class="token string" style="color:#e3116c">"value"</span><span class="token punctuation" style="color:#393A34">}</span><span class="token plain">  </span><span class="token comment" style="color:#999988;font-style:italic"># Optional: pass variables</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">    </span><span class="token punctuation" style="color:#393A34">}</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain"></span><span class="token punctuation" style="color:#393A34">)</span><br></span></code></pre></div></div>
<p>Get started here: <a href="https://docs.litellm.ai/docs/proxy/litellm_prompt_management">LiteLLM Prompt Management Documentation</a></p>
<hr>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="performance--realtime-182-lower-p99-latency">Performance – <code>/realtime</code> 182× Lower p99 Latency<a href="https://docs.litellm.ai/release_notes/v1-80-5#performance--realtime-182-lower-p99-latency" class="hash-link" aria-label="Direct link to performance--realtime-182-lower-p99-latency" title="Direct link to performance--realtime-182-lower-p99-latency">​</a></h3>
<p>This update reduces <code>/realtime</code> latency by removing redundant encodings on the hot path, reusing shared SSL contexts, and caching formatting strings that were being regenerated twice per request despite rarely changing.</p>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="results">Results<a href="https://docs.litellm.ai/release_notes/v1-80-5#results" class="hash-link" aria-label="Direct link to Results" title="Direct link to Results">​</a></h4>
<table><thead><tr><th>Metric</th><th>Before</th><th>After</th><th>Improvement</th></tr></thead><tbody><tr><td>Median latency</td><td>2,200 ms</td><td><strong>59 ms</strong></td><td><strong>−97% (~37× faster)</strong></td></tr><tr><td>p95 latency</td><td>8,500 ms</td><td><strong>67 ms</strong></td><td><strong>−99% (~127× faster)</strong></td></tr><tr><td>p99 latency</td><td>18,000 ms</td><td><strong>99 ms</strong></td><td><strong>−99% (~182× faster)</strong></td></tr><tr><td>Average latency</td><td>3,214 ms</td><td><strong>63 ms</strong></td><td><strong>−98% (~51× faster)</strong></td></tr><tr><td>RPS</td><td>165</td><td><strong>1,207</strong></td><td><strong>+631% (~7.3× increase)</strong></td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="test-setup">Test Setup<a href="https://docs.litellm.ai/release_notes/v1-80-5#test-setup" class="hash-link" aria-label="Direct link to Test Setup" title="Direct link to Test Setup">​</a></h4>
<table><thead><tr><th>Category</th><th>Specification</th></tr></thead><tbody><tr><td><strong>Load Testing</strong></td><td>Locust: 1,000 concurrent users, 500 ramp-up</td></tr><tr><td><strong>System</strong></td><td>4 vCPUs, 8 GB RAM, 4 workers, 4 instances</td></tr><tr><td><strong>Database</strong></td><td>PostgreSQL (Redis unused)</td></tr><tr><td><strong>Configuration</strong></td><td><a href="https://gist.github.com/AlexsanderHamir/420fb44c31c00b4f17a99588637f01ec" target="_blank" rel="noopener noreferrer">config.yaml</a></td></tr><tr><td><strong>Load Script</strong></td><td><a href="https://gist.github.com/AlexsanderHamir/73b83ada21d9b84d4fe09665cf1745f5" target="_blank" rel="noopener noreferrer">no_cache_hits.py</a></td></tr></tbody></table>
<hr>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="model-compare-ui">Model Compare UI<a href="https://docs.litellm.ai/release_notes/v1-80-5#model-compare-ui" class="hash-link" aria-label="Direct link to Model Compare UI" title="Direct link to Model Compare UI">​</a></h3>
<p>New interactive playground UI enables side-by-side comparison of multiple LLM models, making it easy to evaluate and compare model responses.</p>
<p><strong>Features:</strong></p>
<ul>
<li>Compare responses from multiple models in real-time</li>
<li>Side-by-side view with synchronized scrolling</li>
<li>Support for all LiteLLM-supported models</li>
<li>Cost tracking per model</li>
<li>Response time comparison</li>
<li>Pre-configured prompts for quick and easy testing</li>
</ul>
<p><strong>Details:</strong></p>
<ul>
<li>
<p><strong>Parameterization</strong>: Configure API keys, endpoints, models, and model parameters, as well as interaction types (chat completions, embeddings, etc.)</p>
</li>
<li>
<p><strong>Model Comparison</strong>: Compare up to 3 different models simultaneously with side-by-side response views</p>
</li>
<li>
<p><strong>Comparison Metrics</strong>: View detailed comparison information including:</p>
<ul>
<li>Time To First Token</li>
<li>Input / Output / Reasoning Tokens</li>
<li>Total Latency</li>
<li>Cost (if enabled in config)</li>
</ul>
</li>
<li>
<p><strong>Safety Filters</strong>: Configure and test guardrails (safety filters) directly in the playground interface</p>
</li>
</ul>
<p><a href="https://docs.litellm.ai/docs/proxy/model_compare_ui">Get Started with Model Compare</a></p>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-providers-and-endpoints">New Providers and Endpoints<a href="https://docs.litellm.ai/release_notes/v1-80-5#new-providers-and-endpoints" class="hash-link" aria-label="Direct link to New Providers and Endpoints" title="Direct link to New Providers and Endpoints">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="new-providers">New Providers<a href="https://docs.litellm.ai/release_notes/v1-80-5#new-providers" class="hash-link" aria-label="Direct link to New Providers" title="Direct link to New Providers">​</a></h3>
<table><thead><tr><th>Provider</th><th>Supported Endpoints</th><th>Description</th></tr></thead><tbody><tr><td><strong><a href="https://docs.litellm.ai/docs/providers/docker_model_runner">Docker Model Runner</a></strong></td><td><code>/v1/chat/completions</code></td><td>Run LLM models in Docker containers</td></tr></tbody></table>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-80-5#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support">New Model Support<a href="https://docs.litellm.ai/release_notes/v1-80-5#new-model-support" class="hash-link" aria-label="Direct link to New Model Support" title="Direct link to New Model Support">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th><th>Features</th></tr></thead><tbody><tr><td>Azure</td><td><code>azure/gpt-5.1</code></td><td>272K</td><td>$1.38</td><td>$11.00</td><td>Reasoning, vision, PDF input, responses API</td></tr><tr><td>Azure</td><td><code>azure/gpt-5.1-2025-11-13</code></td><td>272K</td><td>$1.38</td><td>$11.00</td><td>Reasoning, vision, PDF input, responses API</td></tr><tr><td>Azure</td><td><code>azure/gpt-5.1-codex</code></td><td>272K</td><td>$1.38</td><td>$11.00</td><td>Responses API, reasoning, vision</td></tr><tr><td>Azure</td><td><code>azure/gpt-5.1-codex-2025-11-13</code></td><td>272K</td><td>$1.38</td><td>$11.00</td><td>Responses API, reasoning, vision</td></tr><tr><td>Azure</td><td><code>azure/gpt-5.1-codex-mini</code></td><td>272K</td><td>$0.275</td><td>$2.20</td><td>Responses API, reasoning, vision</td></tr><tr><td>Azure</td><td><code>azure/gpt-5.1-codex-mini-2025-11-13</code></td><td>272K</td><td>$0.275</td><td>$2.20</td><td>Responses API, reasoning, vision</td></tr><tr><td>Azure EU</td><td><code>azure/eu/gpt-5-2025-08-07</code></td><td>272K</td><td>$1.375</td><td>$11.00</td><td>Reasoning, vision, PDF input</td></tr><tr><td>Azure EU</td><td><code>azure/eu/gpt-5-mini-2025-08-07</code></td><td>272K</td><td>$0.275</td><td>$2.20</td><td>Reasoning, vision, PDF input</td></tr><tr><td>Azure EU</td><td><code>azure/eu/gpt-5-nano-2025-08-07</code></td><td>272K</td><td>$0.055</td><td>$0.44</td><td>Reasoning, vision, PDF input</td></tr><tr><td>Azure EU</td><td><code>azure/eu/gpt-5.1</code></td><td>272K</td><td>$1.38</td><td>$11.00</td><td>Reasoning, vision, PDF input, responses API</td></tr><tr><td>Azure EU</td><td><code>azure/eu/gpt-5.1-codex</code></td><td>272K</td><td>$1.38</td><td>$11.00</td><td>Responses API, reasoning, vision</td></tr><tr><td>Azure EU</td><td><code>azure/eu/gpt-5.1-codex-mini</code></td><td>272K</td><td>$0.275</td><td>$2.20</td><td>Responses API, reasoning, vision</td></tr><tr><td>Gemini</td><td><code>gemini-3-pro-preview</code></td><td>2M</td><td>$1.25</td><td>$5.00</td><td>Reasoning, vision, function calling</td></tr><tr><td>Gemini</td><td><code>gemini-3-pro-image</code></td><td>2M</td><td>$1.25</td><td>$5.00</td><td>Image generation, reasoning</td></tr><tr><td>OpenRouter</td><td><code>openrouter/deepseek/deepseek-v3p1-terminus</code></td><td>164K</td><td>$0.20</td><td>$0.40</td><td>Function calling, reasoning</td></tr><tr><td>OpenRouter</td><td><code>openrouter/moonshot/kimi-k2-instruct</code></td><td>262K</td><td>$0.60</td><td>$2.50</td><td>Function calling, web search</td></tr><tr><td>OpenRouter</td><td><code>openrouter/gemini/gemini-3-pro-preview</code></td><td>2M</td><td>$1.25</td><td>$5.00</td><td>Reasoning, vision, function calling</td></tr><tr><td>XAI</td><td><code>xai/grok-4.1-fast</code></td><td>2M</td><td>$0.20</td><td>$0.50</td><td>Reasoning, function calling</td></tr><tr><td>Together AI</td><td><code>together_ai/z-ai/glm-4.6</code></td><td>203K</td><td>$0.40</td><td>$1.75</td><td>Function calling, reasoning</td></tr><tr><td>Cerebras</td><td><code>cerebras/gpt-oss-120b</code></td><td>131K</td><td>$0.60</td><td>$0.60</td><td>Function calling</td></tr><tr><td>Bedrock</td><td><code>anthropic.claude-sonnet-4-5-20250929-v1:0</code></td><td>200K</td><td>$3.00</td><td>$15.00</td><td>Computer use, reasoning, vision</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-80-5#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Gemini (Google AI Studio + Vertex AI)</a></strong></p>
<ul>
<li>Add Day 0 gemini-3-pro-preview support - <a href="https://github.com/BerriAI/litellm/pull/16719" target="_blank" rel="noopener noreferrer">PR #16719</a></li>
<li>Add support for Gemini 3 Pro Image model - <a href="https://github.com/BerriAI/litellm/pull/16938" target="_blank" rel="noopener noreferrer">PR #16938</a></li>
<li>Add reasoning_content to streaming responses with tools enabled - <a href="https://github.com/BerriAI/litellm/pull/16854" target="_blank" rel="noopener noreferrer">PR #16854</a></li>
<li>Add includeThoughts=True for Gemini 3 reasoning_effort - <a href="https://github.com/BerriAI/litellm/pull/16838" target="_blank" rel="noopener noreferrer">PR #16838</a></li>
<li>Support thought signatures for Gemini 3 in responses API - <a href="https://github.com/BerriAI/litellm/pull/16872" target="_blank" rel="noopener noreferrer">PR #16872</a></li>
<li>Correct wrong system message handling for gemma - <a href="https://github.com/BerriAI/litellm/pull/16767" target="_blank" rel="noopener noreferrer">PR #16767</a></li>
<li>Gemini 3 Pro Image: capture image_tokens and support cost_per_output_image - <a href="https://github.com/BerriAI/litellm/pull/16912" target="_blank" rel="noopener noreferrer">PR #16912</a></li>
<li>Fix missing costs for gemini-2.5-flash-image - <a href="https://github.com/BerriAI/litellm/pull/16882" target="_blank" rel="noopener noreferrer">PR #16882</a></li>
<li>Gemini 3 thought signatures in tool call id - <a href="https://github.com/BerriAI/litellm/pull/16895" target="_blank" rel="noopener noreferrer">PR #16895</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/azure">Azure</a></strong></p>
<ul>
<li>Add azure gpt-5.1 models - <a href="https://github.com/BerriAI/litellm/pull/16817" target="_blank" rel="noopener noreferrer">PR #16817</a></li>
<li>Add Azure models 2025 11 to cost maps - <a href="https://github.com/BerriAI/litellm/pull/16762" target="_blank" rel="noopener noreferrer">PR #16762</a></li>
<li>Update Azure Pricing - <a href="https://github.com/BerriAI/litellm/pull/16371" target="_blank" rel="noopener noreferrer">PR #16371</a></li>
<li>Add SSML Support for Azure Text-to-Speech (AVA) - <a href="https://github.com/BerriAI/litellm/pull/16747" target="_blank" rel="noopener noreferrer">PR #16747</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI</a></strong></p>
<ul>
<li>Support GPT-5.1 reasoning.effort='none' in proxy - <a href="https://github.com/BerriAI/litellm/pull/16745" target="_blank" rel="noopener noreferrer">PR #16745</a></li>
<li>Add gpt-5.1-codex and gpt-5.1-codex-mini models to documentation - <a href="https://github.com/BerriAI/litellm/pull/16735" target="_blank" rel="noopener noreferrer">PR #16735</a></li>
<li>Inherit BaseVideoConfig to enable async content response for OpenAI video - <a href="https://github.com/BerriAI/litellm/pull/16708" target="_blank" rel="noopener noreferrer">PR #16708</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Add support for <code>strict</code> parameter in Anthropic tool schemas - <a href="https://github.com/BerriAI/litellm/pull/16725" target="_blank" rel="noopener noreferrer">PR #16725</a></li>
<li>Add image as url support to anthropic - <a href="https://github.com/BerriAI/litellm/pull/16868" target="_blank" rel="noopener noreferrer">PR #16868</a></li>
<li>Add thought signature support to v1/messages api - <a href="https://github.com/BerriAI/litellm/pull/16812" target="_blank" rel="noopener noreferrer">PR #16812</a></li>
<li>Anthropic - support Structured Outputs <code>output_format</code> for Claude 4.5 sonnet and Opus 4.1 - <a href="https://github.com/BerriAI/litellm/pull/16949" target="_blank" rel="noopener noreferrer">PR #16949</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong></p>
<ul>
<li>Haiku 4.5 correct Bedrock configs - <a href="https://github.com/BerriAI/litellm/pull/16732" target="_blank" rel="noopener noreferrer">PR #16732</a></li>
<li>Ensure consistent chunk IDs in Bedrock streaming responses - <a href="https://github.com/BerriAI/litellm/pull/16596" target="_blank" rel="noopener noreferrer">PR #16596</a></li>
<li>Add Claude 4.5 to US Gov Cloud - <a href="https://github.com/BerriAI/litellm/pull/16957" target="_blank" rel="noopener noreferrer">PR #16957</a></li>
<li>Fix images being dropped from tool results for bedrock - <a href="https://github.com/BerriAI/litellm/pull/16492" target="_blank" rel="noopener noreferrer">PR #16492</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vertex">Vertex AI</a></strong></p>
<ul>
<li>Add Vertex AI Image Edit Support - <a href="https://github.com/BerriAI/litellm/pull/16828" target="_blank" rel="noopener noreferrer">PR #16828</a></li>
<li>Update veo 3 pricing and add prod models - <a href="https://github.com/BerriAI/litellm/pull/16781" target="_blank" rel="noopener noreferrer">PR #16781</a></li>
<li>Fix Video download for veo3 - <a href="https://github.com/BerriAI/litellm/pull/16875" target="_blank" rel="noopener noreferrer">PR #16875</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/snowflake">Snowflake</a></strong></p>
<ul>
<li>Snowflake provider support: added embeddings, PAT, account_id - <a href="https://github.com/BerriAI/litellm/pull/15727" target="_blank" rel="noopener noreferrer">PR #15727</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/oci">OCI</a></strong></p>
<ul>
<li>Add oci_endpoint_id Parameter for OCI Dedicated Endpoints - <a href="https://github.com/BerriAI/litellm/pull/16723" target="_blank" rel="noopener noreferrer">PR #16723</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/xai">XAI</a></strong></p>
<ul>
<li>Add support for Grok 4.1 Fast models - <a href="https://github.com/BerriAI/litellm/pull/16936" target="_blank" rel="noopener noreferrer">PR #16936</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/togetherai">Together AI</a></strong></p>
<ul>
<li>Add GLM 4.6 from together.ai - <a href="https://github.com/BerriAI/litellm/pull/16942" target="_blank" rel="noopener noreferrer">PR #16942</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/cerebras">Cerebras</a></strong></p>
<ul>
<li>Fix Cerebras GPT-OSS-120B model name - <a href="https://github.com/BerriAI/litellm/pull/16939" target="_blank" rel="noopener noreferrer">PR #16939</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-80-5#bug-fixes" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI</a></strong></p>
<ul>
<li>Fix for 16863 - openai conversion from responses to completions - <a href="https://github.com/BerriAI/litellm/pull/16864" target="_blank" rel="noopener noreferrer">PR #16864</a></li>
<li>Revert "Make all gpt-5 and reasoning models to responses by default" - <a href="https://github.com/BerriAI/litellm/pull/16849" target="_blank" rel="noopener noreferrer">PR #16849</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Get custom_llm_provider from query param - <a href="https://github.com/BerriAI/litellm/pull/16731" target="_blank" rel="noopener noreferrer">PR #16731</a></li>
<li>Fix optional param mapping - <a href="https://github.com/BerriAI/litellm/pull/16852" target="_blank" rel="noopener noreferrer">PR #16852</a></li>
<li>Add None check for litellm_params - <a href="https://github.com/BerriAI/litellm/pull/16754" target="_blank" rel="noopener noreferrer">PR #16754</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-80-5#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-80-5#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/response_api">Responses API</a></strong></p>
<ul>
<li>Add Responses API support for gpt-5.1-codex model - <a href="https://github.com/BerriAI/litellm/pull/16845" target="_blank" rel="noopener noreferrer">PR #16845</a></li>
<li>Add managed files support for responses API - <a href="https://github.com/BerriAI/litellm/pull/16733" target="_blank" rel="noopener noreferrer">PR #16733</a></li>
<li>Add extra_body support for response supported api params from chat completion - <a href="https://github.com/BerriAI/litellm/pull/16765" target="_blank" rel="noopener noreferrer">PR #16765</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/batches">Batch API</a></strong></p>
<ul>
<li>Support /delete for files + support /cancel for batches - <a href="https://github.com/BerriAI/litellm/pull/16387" target="_blank" rel="noopener noreferrer">PR #16387</a></li>
<li>Add config based routing support for batches and files - <a href="https://github.com/BerriAI/litellm/pull/16872" target="_blank" rel="noopener noreferrer">PR #16872</a></li>
<li>Populate spend_logs_metadata in batch and files endpoints - <a href="https://github.com/BerriAI/litellm/pull/16921" target="_blank" rel="noopener noreferrer">PR #16921</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/search">Search APIs</a></strong></p>
<ul>
<li>Search APIs - error in firecrawl-search "Invalid request body" - <a href="https://github.com/BerriAI/litellm/pull/16943" target="_blank" rel="noopener noreferrer">PR #16943</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/vector_stores">Vector Stores</a></strong></p>
<ul>
<li>Fix vector store create issue - <a href="https://github.com/BerriAI/litellm/pull/16804" target="_blank" rel="noopener noreferrer">PR #16804</a></li>
<li>Team vector-store permissions now respected for key access - <a href="https://github.com/BerriAI/litellm/pull/16639" target="_blank" rel="noopener noreferrer">PR #16639</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/audio_transcription">Audio Transcription</a></strong></p>
<ul>
<li>Fix audio transcription cost tracking - <a href="https://github.com/BerriAI/litellm/pull/16478" target="_blank" rel="noopener noreferrer">PR #16478</a></li>
<li>Add missing shared_sessions to audio/transcriptions - <a href="https://github.com/BerriAI/litellm/pull/16858" target="_blank" rel="noopener noreferrer">PR #16858</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/video_generation">Video Generation API</a></strong></p>
<ul>
<li>Fix videos tagging - <a href="https://github.com/BerriAI/litellm/pull/16770" target="_blank" rel="noopener noreferrer">PR #16770</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs">Bugs<a href="https://docs.litellm.ai/release_notes/v1-80-5#bugs" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>General</strong>
<ul>
<li>Responses API cost tracking with custom deployment names - <a href="https://github.com/BerriAI/litellm/pull/16778" target="_blank" rel="noopener noreferrer">PR #16778</a></li>
<li>Trim logged response strings in spend-logs - <a href="https://github.com/BerriAI/litellm/pull/16654" target="_blank" rel="noopener noreferrer">PR #16654</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-80-5#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-80-5#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong>Proxy CLI Auth</strong></p>
<ul>
<li>Allow using JWTs for signing in with Proxy CLI - <a href="https://github.com/BerriAI/litellm/pull/16756" target="_blank" rel="noopener noreferrer">PR #16756</a></li>
</ul>
</li>
<li>
<p><strong>Virtual Keys</strong></p>
<ul>
<li>Fix Key Model Alias Not Working - <a href="https://github.com/BerriAI/litellm/pull/16896" target="_blank" rel="noopener noreferrer">PR #16896</a></li>
</ul>
</li>
<li>
<p><strong>Models + Endpoints</strong></p>
<ul>
<li>Add additional model settings to chat models in test key - <a href="https://github.com/BerriAI/litellm/pull/16793" target="_blank" rel="noopener noreferrer">PR #16793</a></li>
<li>Deactivate delete button on model table for config models - <a href="https://github.com/BerriAI/litellm/pull/16787" target="_blank" rel="noopener noreferrer">PR #16787</a></li>
<li>Change Public Model Hub to use proxyBaseUrl - <a href="https://github.com/BerriAI/litellm/pull/16892" target="_blank" rel="noopener noreferrer">PR #16892</a></li>
<li>Add JSON Viewer to request/response panel - <a href="https://github.com/BerriAI/litellm/pull/16687" target="_blank" rel="noopener noreferrer">PR #16687</a></li>
<li>Standarize icon images - <a href="https://github.com/BerriAI/litellm/pull/16837" target="_blank" rel="noopener noreferrer">PR #16837</a></li>
</ul>
</li>
<li>
<p><strong>Teams</strong></p>
<ul>
<li>Teams table empty state - <a href="https://github.com/BerriAI/litellm/pull/16738" target="_blank" rel="noopener noreferrer">PR #16738</a></li>
</ul>
</li>
<li>
<p><strong>Fallbacks</strong></p>
<ul>
<li>Fallbacks icon button tooltips and delete with friction - <a href="https://github.com/BerriAI/litellm/pull/16737" target="_blank" rel="noopener noreferrer">PR #16737</a></li>
</ul>
</li>
<li>
<p><strong>MCP Servers</strong></p>
<ul>
<li>Delete user and MCP Server Modal, MCP Table Tooltips - <a href="https://github.com/BerriAI/litellm/pull/16751" target="_blank" rel="noopener noreferrer">PR #16751</a></li>
</ul>
</li>
<li>
<p><strong>Callbacks</strong></p>
<ul>
<li>Expose backend endpoint for callbacks settings - <a href="https://github.com/BerriAI/litellm/pull/16698" target="_blank" rel="noopener noreferrer">PR #16698</a></li>
<li>Edit add callbacks route to use data from backend - <a href="https://github.com/BerriAI/litellm/pull/16699" target="_blank" rel="noopener noreferrer">PR #16699</a></li>
</ul>
</li>
<li>
<p><strong>Usage &amp; Analytics</strong></p>
<ul>
<li>Allow partial matches for user ID in User Table - <a href="https://github.com/BerriAI/litellm/pull/16952" target="_blank" rel="noopener noreferrer">PR #16952</a></li>
</ul>
</li>
<li>
<p><strong>General UI</strong></p>
<ul>
<li>Allow setting base_url in API reference docs - <a href="https://github.com/BerriAI/litellm/pull/16674" target="_blank" rel="noopener noreferrer">PR #16674</a></li>
<li>Change /public fields to honor server root path - <a href="https://github.com/BerriAI/litellm/pull/16930" target="_blank" rel="noopener noreferrer">PR #16930</a></li>
<li>Correct ui build - <a href="https://github.com/BerriAI/litellm/pull/16702" target="_blank" rel="noopener noreferrer">PR #16702</a></li>
<li>Enable automatic dark/light mode based on system preference - <a href="https://github.com/BerriAI/litellm/pull/16748" target="_blank" rel="noopener noreferrer">PR #16748</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs-1">Bugs<a href="https://docs.litellm.ai/release_notes/v1-80-5#bugs-1" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li>
<p><strong>UI Fixes</strong></p>
<ul>
<li>Fix flaky tests due to antd Notification Manager - <a href="https://github.com/BerriAI/litellm/pull/16740" target="_blank" rel="noopener noreferrer">PR #16740</a></li>
<li>Fix UI MCP Tool Test Regression - <a href="https://github.com/BerriAI/litellm/pull/16695" target="_blank" rel="noopener noreferrer">PR #16695</a></li>
<li>Fix edit logging settings not appearing - <a href="https://github.com/BerriAI/litellm/pull/16798" target="_blank" rel="noopener noreferrer">PR #16798</a></li>
<li>Add css to truncate long request ids in request viewer - <a href="https://github.com/BerriAI/litellm/pull/16665" target="_blank" rel="noopener noreferrer">PR #16665</a></li>
<li>Remove azure/ prefix in Placeholder for Azure in Add Model - <a href="https://github.com/BerriAI/litellm/pull/16597" target="_blank" rel="noopener noreferrer">PR #16597</a></li>
<li>Remove UI Session Token from user/info return - <a href="https://github.com/BerriAI/litellm/pull/16851" target="_blank" rel="noopener noreferrer">PR #16851</a></li>
<li>Remove console logs and errors from model tab - <a href="https://github.com/BerriAI/litellm/pull/16455" target="_blank" rel="noopener noreferrer">PR #16455</a></li>
<li>Change Bulk Invite User Roles to Match Backend - <a href="https://github.com/BerriAI/litellm/pull/16906" target="_blank" rel="noopener noreferrer">PR #16906</a></li>
<li>Mock Tremor's Tooltip to Fix Flaky UI Tests - <a href="https://github.com/BerriAI/litellm/pull/16786" target="_blank" rel="noopener noreferrer">PR #16786</a></li>
<li>Fix e2e ui playwright test - <a href="https://github.com/BerriAI/litellm/pull/16799" target="_blank" rel="noopener noreferrer">PR #16799</a></li>
<li>Fix Tests in CI/CD - <a href="https://github.com/BerriAI/litellm/pull/16972" target="_blank" rel="noopener noreferrer">PR #16972</a></li>
</ul>
</li>
<li>
<p><strong>SSO</strong></p>
<ul>
<li>Ensure <code>role</code> from SSO provider is used when a user is inserted onto LiteLLM - <a href="https://github.com/BerriAI/litellm/pull/16794" target="_blank" rel="noopener noreferrer">PR #16794</a></li>
<li>Docs - SSO - Manage User Roles via Azure App Roles - <a href="https://github.com/BerriAI/litellm/pull/16796" target="_blank" rel="noopener noreferrer">PR #16796</a></li>
</ul>
</li>
<li>
<p><strong>Auth</strong></p>
<ul>
<li>Ensure Team Tags works when using JWT Auth - <a href="https://github.com/BerriAI/litellm/pull/16797" target="_blank" rel="noopener noreferrer">PR #16797</a></li>
<li>Fix key never expires - <a href="https://github.com/BerriAI/litellm/pull/16692" target="_blank" rel="noopener noreferrer">PR #16692</a></li>
</ul>
</li>
<li>
<p><strong>Swagger UI</strong></p>
<ul>
<li>Fixes Swagger UI resolver errors for chat completion endpoints caused by Pydantic v2 <code>$defs</code> not being properly exposed in the OpenAPI schema - <a href="https://github.com/BerriAI/litellm/pull/16784" target="_blank" rel="noopener noreferrer">PR #16784</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="ai-integrations">AI Integrations<a href="https://docs.litellm.ai/release_notes/v1-80-5#ai-integrations" class="hash-link" aria-label="Direct link to AI Integrations" title="Direct link to AI Integrations">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="logging">Logging<a href="https://docs.litellm.ai/release_notes/v1-80-5#logging" class="hash-link" aria-label="Direct link to Logging" title="Direct link to Logging">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/observability/arize_phoenix">Arize Phoenix</a></strong></p>
<ul>
<li>Fix arize phoenix logging - <a href="https://github.com/BerriAI/litellm/pull/16301" target="_blank" rel="noopener noreferrer">PR #16301</a></li>
<li>Arize Phoenix - root span logging - <a href="https://github.com/BerriAI/litellm/pull/16949" target="_blank" rel="noopener noreferrer">PR #16949</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#langfuse">Langfuse</a></strong></p>
<ul>
<li>Filter secret fields form Langfuse - <a href="https://github.com/BerriAI/litellm/pull/16842" target="_blank" rel="noopener noreferrer">PR #16842</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Exclude litellm_credential_name from Sensitive Data Masker (Updated) - <a href="https://github.com/BerriAI/litellm/pull/16958" target="_blank" rel="noopener noreferrer">PR #16958</a></li>
<li>Allow admins to disable, dynamic callback controls - <a href="https://github.com/BerriAI/litellm/pull/16750" target="_blank" rel="noopener noreferrer">PR #16750</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails">Guardrails<a href="https://docs.litellm.ai/release_notes/v1-80-5#guardrails" class="hash-link" aria-label="Direct link to Guardrails" title="Direct link to Guardrails">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">IBM Guardrails</a></strong></p>
<ul>
<li>Fix IBM Guardrails optional params, add extra_headers field - <a href="https://github.com/BerriAI/litellm/pull/16771" target="_blank" rel="noopener noreferrer">PR #16771</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">Noma Guardrail</a></strong></p>
<ul>
<li>Use LiteLLM key alias as fallback Noma applicationId in NomaGuardrail - <a href="https://github.com/BerriAI/litellm/pull/16832" target="_blank" rel="noopener noreferrer">PR #16832</a></li>
<li>Allow custom violation message for tool-permission guardrail - <a href="https://github.com/BerriAI/litellm/pull/16916" target="_blank" rel="noopener noreferrer">PR #16916</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">Grayswan Guardrail</a></strong></p>
<ul>
<li>Grayswan guardrail passthrough on flagged - <a href="https://github.com/BerriAI/litellm/pull/16891" target="_blank" rel="noopener noreferrer">PR #16891</a></li>
</ul>
</li>
<li>
<p><strong>General Guardrails</strong></p>
<ul>
<li>Fix prompt injection not working - <a href="https://github.com/BerriAI/litellm/pull/16701" target="_blank" rel="noopener noreferrer">PR #16701</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="prompt-management-1">Prompt Management<a href="https://docs.litellm.ai/release_notes/v1-80-5#prompt-management-1" class="hash-link" aria-label="Direct link to Prompt Management" title="Direct link to Prompt Management">​</a></h3>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/prompt_management">Prompt Management</a></strong>
<ul>
<li>Allow specifying just prompt_id in a request to a model - <a href="https://github.com/BerriAI/litellm/pull/16834" target="_blank" rel="noopener noreferrer">PR #16834</a></li>
<li>Add support for versioning prompts - <a href="https://github.com/BerriAI/litellm/pull/16836" target="_blank" rel="noopener noreferrer">PR #16836</a></li>
<li>Allow storing prompt version in DB - <a href="https://github.com/BerriAI/litellm/pull/16848" target="_blank" rel="noopener noreferrer">PR #16848</a></li>
<li>Add UI for editing the prompts - <a href="https://github.com/BerriAI/litellm/pull/16853" target="_blank" rel="noopener noreferrer">PR #16853</a></li>
<li>Allow testing prompts with Chat UI - <a href="https://github.com/BerriAI/litellm/pull/16898" target="_blank" rel="noopener noreferrer">PR #16898</a></li>
<li>Allow viewing version history - <a href="https://github.com/BerriAI/litellm/pull/16901" target="_blank" rel="noopener noreferrer">PR #16901</a></li>
<li>Allow specifying prompt version in code - <a href="https://github.com/BerriAI/litellm/pull/16929" target="_blank" rel="noopener noreferrer">PR #16929</a></li>
<li>UI, allow seeing model, prompt id for Prompt - <a href="https://github.com/BerriAI/litellm/pull/16932" target="_blank" rel="noopener noreferrer">PR #16932</a></li>
<li>Show "get code" section for prompt management + minor polish of showing version history - <a href="https://github.com/BerriAI/litellm/pull/16941" target="_blank" rel="noopener noreferrer">PR #16941</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="secret-managers">Secret Managers<a href="https://docs.litellm.ai/release_notes/v1-80-5#secret-managers" class="hash-link" aria-label="Direct link to Secret Managers" title="Direct link to Secret Managers">​</a></h3>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/secret_managers">AWS Secrets Manager</a></strong>
<ul>
<li>Adds IAM role assumption support for AWS Secret Manager - <a href="https://github.com/BerriAI/litellm/pull/16887" target="_blank" rel="noopener noreferrer">PR #16887</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-gateway">MCP Gateway<a href="https://docs.litellm.ai/release_notes/v1-80-5#mcp-gateway" class="hash-link" aria-label="Direct link to MCP Gateway" title="Direct link to MCP Gateway">​</a></h2>
<ul>
<li><strong>MCP Hub</strong> - Publish/discover MCP Servers within a company - <a href="https://github.com/BerriAI/litellm/pull/16857" target="_blank" rel="noopener noreferrer">PR #16857</a></li>
<li><strong>MCP Resources</strong> - MCP resources support - <a href="https://github.com/BerriAI/litellm/pull/16800" target="_blank" rel="noopener noreferrer">PR #16800</a></li>
<li><strong>MCP OAuth</strong> - Docs - mcp oauth flow details - <a href="https://github.com/BerriAI/litellm/pull/16742" target="_blank" rel="noopener noreferrer">PR #16742</a></li>
<li><strong>MCP Lifecycle</strong> - Drop MCPClient.connect and use run_with_session lifecycle - <a href="https://github.com/BerriAI/litellm/pull/16696" target="_blank" rel="noopener noreferrer">PR #16696</a></li>
<li><strong>MCP Server IDs</strong> - Add mcp server ids - <a href="https://github.com/BerriAI/litellm/pull/16904" target="_blank" rel="noopener noreferrer">PR #16904</a></li>
<li><strong>MCP URL Format</strong> - Fix mcp url format - <a href="https://github.com/BerriAI/litellm/pull/16940" target="_blank" rel="noopener noreferrer">PR #16940</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements">Performance / Loadbalancing / Reliability improvements<a href="https://docs.litellm.ai/release_notes/v1-80-5#performance--loadbalancing--reliability-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements" title="Direct link to Performance / Loadbalancing / Reliability improvements">​</a></h2>
<ul>
<li><strong>Realtime Endpoint Performance</strong> - Fix bottlenecks degrading realtime endpoint performance - <a href="https://github.com/BerriAI/litellm/pull/16670" target="_blank" rel="noopener noreferrer">PR #16670</a></li>
<li><strong>SSL Context Caching</strong> - Cache SSL contexts to prevent excessive memory allocation - <a href="https://github.com/BerriAI/litellm/pull/16955" target="_blank" rel="noopener noreferrer">PR #16955</a></li>
<li><strong>Cache Optimization</strong> - Fix cache cooldown key generation - <a href="https://github.com/BerriAI/litellm/pull/16954" target="_blank" rel="noopener noreferrer">PR #16954</a></li>
<li><strong>Router Cache</strong> - Fix routing for requests with same cacheable prefix but different user messages - <a href="https://github.com/BerriAI/litellm/pull/16951" target="_blank" rel="noopener noreferrer">PR #16951</a></li>
<li><strong>Redis Event Loop</strong> - Fix redis event loop closed at first call - <a href="https://github.com/BerriAI/litellm/pull/16913" target="_blank" rel="noopener noreferrer">PR #16913</a></li>
<li><strong>Dependency Management</strong> - Upgrade pydantic to version 2.11.0 - <a href="https://github.com/BerriAI/litellm/pull/16909" target="_blank" rel="noopener noreferrer">PR #16909</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates">Documentation Updates<a href="https://docs.litellm.ai/release_notes/v1-80-5#documentation-updates" class="hash-link" aria-label="Direct link to Documentation Updates" title="Direct link to Documentation Updates">​</a></h2>
<ul>
<li>
<p><strong>Provider Documentation</strong></p>
<ul>
<li>Add missing details to benchmark comparison - <a href="https://github.com/BerriAI/litellm/pull/16690" target="_blank" rel="noopener noreferrer">PR #16690</a></li>
<li>Fix anthropic pass-through endpoint - <a href="https://github.com/BerriAI/litellm/pull/16883" target="_blank" rel="noopener noreferrer">PR #16883</a></li>
<li>Cleanup repo and improve AI docs - <a href="https://github.com/BerriAI/litellm/pull/16775" target="_blank" rel="noopener noreferrer">PR #16775</a></li>
</ul>
</li>
<li>
<p><strong>API Documentation</strong></p>
<ul>
<li>Add docs related to openai metadata - <a href="https://github.com/BerriAI/litellm/pull/16872" target="_blank" rel="noopener noreferrer">PR #16872</a></li>
<li>Update docs with all supported endpoints and cost tracking - <a href="https://github.com/BerriAI/litellm/pull/16872" target="_blank" rel="noopener noreferrer">PR #16872</a></li>
</ul>
</li>
<li>
<p><strong>General Documentation</strong></p>
<ul>
<li>Add mini-swe-agent to Projects built on LiteLLM - <a href="https://github.com/BerriAI/litellm/pull/16971" target="_blank" rel="noopener noreferrer">PR #16971</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="infrastructure--cicd">Infrastructure / CI/CD<a href="https://docs.litellm.ai/release_notes/v1-80-5#infrastructure--cicd" class="hash-link" aria-label="Direct link to Infrastructure / CI/CD" title="Direct link to Infrastructure / CI/CD">​</a></h2>
<ul>
<li>
<p><strong>UI Testing</strong></p>
<ul>
<li>Break e2e_ui_testing into build, unit, and e2e steps - <a href="https://github.com/BerriAI/litellm/pull/16783" target="_blank" rel="noopener noreferrer">PR #16783</a></li>
<li>Building UI for Testing - <a href="https://github.com/BerriAI/litellm/pull/16968" target="_blank" rel="noopener noreferrer">PR #16968</a></li>
<li>CI/CD Fixes - <a href="https://github.com/BerriAI/litellm/pull/16937" target="_blank" rel="noopener noreferrer">PR #16937</a></li>
</ul>
</li>
<li>
<p><strong>Dependency Management</strong></p>
<ul>
<li>Bump js-yaml from 3.14.1 to 3.14.2 in /tests/proxy_admin_ui_tests/ui_unit_tests - <a href="https://github.com/BerriAI/litellm/pull/16755" target="_blank" rel="noopener noreferrer">PR #16755</a></li>
<li>Bump js-yaml from 3.14.1 to 3.14.2 - <a href="https://github.com/BerriAI/litellm/pull/16802" target="_blank" rel="noopener noreferrer">PR #16802</a></li>
</ul>
</li>
<li>
<p><strong>Migration</strong></p>
<ul>
<li>Migration job labels - <a href="https://github.com/BerriAI/litellm/pull/16831" target="_blank" rel="noopener noreferrer">PR #16831</a></li>
</ul>
</li>
<li>
<p><strong>Config</strong></p>
<ul>
<li>This yaml actually works - <a href="https://github.com/BerriAI/litellm/pull/16757" target="_blank" rel="noopener noreferrer">PR #16757</a></li>
</ul>
</li>
<li>
<p><strong>Release Notes</strong></p>
<ul>
<li>Add perf improvements on embeddings to release notes - <a href="https://github.com/BerriAI/litellm/pull/16697" target="_blank" rel="noopener noreferrer">PR #16697</a></li>
<li>Docs - v1.80.0 - <a href="https://github.com/BerriAI/litellm/pull/16694" target="_blank" rel="noopener noreferrer">PR #16694</a></li>
</ul>
</li>
<li>
<p><strong>Investigation</strong></p>
<ul>
<li>Investigate issue root cause - <a href="https://github.com/BerriAI/litellm/pull/16859" target="_blank" rel="noopener noreferrer">PR #16859</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-80-5#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@mattmorgis made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16371" target="_blank" rel="noopener noreferrer">PR #16371</a></li>
<li>@mmandic-coatue made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16732" target="_blank" rel="noopener noreferrer">PR #16732</a></li>
<li>@Bradley-Butcher made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16725" target="_blank" rel="noopener noreferrer">PR #16725</a></li>
<li>@BenjaminLevy made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16757" target="_blank" rel="noopener noreferrer">PR #16757</a></li>
<li>@CatBraaain made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16767" target="_blank" rel="noopener noreferrer">PR #16767</a></li>
<li>@tushar8408 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16831" target="_blank" rel="noopener noreferrer">PR #16831</a></li>
<li>@nbsp1221 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16845" target="_blank" rel="noopener noreferrer">PR #16845</a></li>
<li>@idola9 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16832" target="_blank" rel="noopener noreferrer">PR #16832</a></li>
<li>@nkukard made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16864" target="_blank" rel="noopener noreferrer">PR #16864</a></li>
<li>@alhuang10 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16852" target="_blank" rel="noopener noreferrer">PR #16852</a></li>
<li>@sebslight made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16838" target="_blank" rel="noopener noreferrer">PR #16838</a></li>
<li>@TsurumaruTsuyoshi made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16905" target="_blank" rel="noopener noreferrer">PR #16905</a></li>
<li>@cyberjunk made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16492" target="_blank" rel="noopener noreferrer">PR #16492</a></li>
<li>@colinlin-stripe made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16895" target="_blank" rel="noopener noreferrer">PR #16895</a></li>
<li>@sureshdsk made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16883" target="_blank" rel="noopener noreferrer">PR #16883</a></li>
<li>@eiliyaabedini made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16875" target="_blank" rel="noopener noreferrer">PR #16875</a></li>
<li>@justin-tahara made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16957" target="_blank" rel="noopener noreferrer">PR #16957</a></li>
<li>@wangsoft made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16913" target="_blank" rel="noopener noreferrer">PR #16913</a></li>
<li>@dsduenas made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16891" target="_blank" rel="noopener noreferrer">PR #16891</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="known-issues">Known Issues<a href="https://docs.litellm.ai/release_notes/v1-80-5#known-issues" class="hash-link" aria-label="Direct link to Known Issues" title="Direct link to Known Issues">​</a></h2>
<ul>
<li><code>/audit</code> and <code>/user/available_users</code> routes return 404. Fixed in <a href="https://github.com/BerriAI/litellm/pull/17337" target="_blank" rel="noopener noreferrer">PR #17337</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog">Full Changelog<a href="https://docs.litellm.ai/release_notes/v1-80-5#full-changelog" class="hash-link" aria-label="Direct link to Full Changelog" title="Direct link to Full Changelog">​</a></h2>
<p><strong><a href="https://github.com/BerriAI/litellm/compare/v1.80.0-nightly...v1.80.5.rc.2" target="_blank" rel="noopener noreferrer">View complete changelog on GitHub</a></strong></p>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[v1.80.0-stable - Introducing Agent Hub: Register, Publish, and Share Agents]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-80-0</id>
        <link href="https://docs.litellm.ai/release_notes/v1-80-0"/>
        <updated>2025-11-15T10:00:00.000Z</updated>
        <summary type="html"><![CDATA[Deploy this version]]></summary>
        <content type="html"><![CDATA[<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-80-0#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">docker.litellm.ai/berriai/litellm:v1.80.0-stable</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.80.0</span><br></span></code></pre></div></div></div></div></div>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-80-0#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<ul>
<li><strong>🆕 Agent Hub Support</strong> - Register and make agents public for your organization</li>
<li><strong>RunwayML Provider</strong> - Complete video generation, image generation, and text-to-speech support</li>
<li><strong>GPT-5.1 Family Support</strong> - Day-0 support for OpenAI's latest GPT-5.1 and GPT-5.1-Codex models</li>
<li><strong>Prometheus OSS</strong> - Prometheus metrics now available in open-source version</li>
<li><strong>Vector Store Files API</strong> - Complete OpenAI-compatible Vector Store Files API with full CRUD operations</li>
<li><strong>Embeddings Performance</strong> - O(1) lookup optimization for router embeddings with shared sessions</li>
</ul>
<hr>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="agent-hub">Agent Hub<a href="https://docs.litellm.ai/release_notes/v1-80-0#agent-hub" class="hash-link" aria-label="Direct link to Agent Hub" title="Direct link to Agent Hub">​</a></h3>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAFCAYAAAB8ZH1oAAAACXBIWXMAABYlAAAWJQFJUiTwAAAAb0lEQVR4nE2N0QqFIQiDe//XDIKgiyBK03QHO/zQ4GPgZEu1VuScMecEEWERQVRh5+A8pNYaSinovWOMgbUWmDfMDJ/cHSm+4xiuqreVmG/4kvbe+BCRO81b4P5vCo4ZUoQvixiTBMccIgrVWHT8AADXw6viLsfXAAAAAElFTkSuQmCC&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="334"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/agent_hub_clean.5715779.640.png" srcset="/assets/ideal-img/agent_hub_clean.5715779.640.png 640w,/assets/ideal-img/agent_hub_clean.59a4bbe.1920.png 1920w" width="640" height="334"></noscript></div>
<p>This release adds support for registering and making agents public for your organization. This is great for <strong>Proxy Admins</strong> who want a central place to make agents built in their organization, discoverable to their users.</p>
<p>Here's the flow:</p>
<ol>
<li>Add agent to litellm.</li>
<li>Make it public.</li>
<li>Allow anyone to discover it on the public AI Hub page.</li>
</ol>
<p><a href="https://docs.litellm.ai/docs/proxy/ai_hub"><strong>Get Started with Agent Hub</strong></a></p>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="performance--embeddings-13-lower-p95-latency">Performance – <code>/embeddings</code> 13× Lower p95 Latency<a href="https://docs.litellm.ai/release_notes/v1-80-0#performance--embeddings-13-lower-p95-latency" class="hash-link" aria-label="Direct link to performance--embeddings-13-lower-p95-latency" title="Direct link to performance--embeddings-13-lower-p95-latency">​</a></h3>
<p>This update significantly improves <code>/embeddings</code> latency by routing it through the same optimized pipeline as <code>/chat/completions</code>, benefiting from all previously applied networking optimizations.</p>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="results">Results<a href="https://docs.litellm.ai/release_notes/v1-80-0#results" class="hash-link" aria-label="Direct link to Results" title="Direct link to Results">​</a></h3>
<table><thead><tr><th>Metric</th><th>Before</th><th>After</th><th>Improvement</th></tr></thead><tbody><tr><td>p95 latency</td><td>5,700 ms</td><td><strong>430 ms</strong></td><td>−92% (~13× faster)**</td></tr><tr><td>p99 latency</td><td>7,200 ms</td><td><strong>780 ms</strong></td><td>−89%</td></tr><tr><td>Average latency</td><td>844 ms</td><td><strong>262 ms</strong></td><td>−69%</td></tr><tr><td>Median latency</td><td>290 ms</td><td><strong>230 ms</strong></td><td>−21%</td></tr><tr><td>RPS</td><td>1,216.7</td><td><strong>1,219.7</strong></td><td><strong>+0.25%</strong></td></tr></tbody></table>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="test-setup">Test Setup<a href="https://docs.litellm.ai/release_notes/v1-80-0#test-setup" class="hash-link" aria-label="Direct link to Test Setup" title="Direct link to Test Setup">​</a></h3>
<table><thead><tr><th>Category</th><th>Specification</th></tr></thead><tbody><tr><td><strong>Load Testing</strong></td><td>Locust: 1,000 concurrent users, 500 ramp-up</td></tr><tr><td><strong>System</strong></td><td>4 vCPUs, 8 GB RAM, 4 workers, 4 instances</td></tr><tr><td><strong>Database</strong></td><td>PostgreSQL (Redis unused)</td></tr><tr><td><strong>Configuration</strong></td><td><a href="https://gist.github.com/AlexsanderHamir/550791675fd752befcac6a9e44024652" target="_blank" rel="noopener noreferrer">config.yaml</a></td></tr><tr><td><strong>Load Script</strong></td><td><a href="https://gist.github.com/AlexsanderHamir/99d673bf74cdd81fd39f59fa9048f2e8" target="_blank" rel="noopener noreferrer">no_cache_hits.py</a></td></tr></tbody></table>
<hr>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="-runwayml">🆕 RunwayML<a href="https://docs.litellm.ai/release_notes/v1-80-0#-runwayml" class="hash-link" aria-label="Direct link to 🆕 RunwayML" title="Direct link to 🆕 RunwayML">​</a></h3>
<p>Complete integration for RunwayML's Gen-4 family of models, supporting video generation, image generation, and text-to-speech.</p>
<p><strong>Supported Endpoints:</strong></p>
<ul>
<li><code>/v1/videos</code> - Video generation (Gen-4 Turbo, Gen-4 Aleph, Gen-3A Turbo)</li>
<li><code>/v1/images/generations</code> - Image generation (Gen-4 Image, Gen-4 Image Turbo)</li>
<li><code>/v1/audio/speech</code> - Text-to-speech (ElevenLabs Multilingual v2)</li>
</ul>
<p><strong>Quick Start:</strong></p>
<div class="language-bash codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">Generate Video with RunwayML</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-bash codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv codeBlockLinesWithNumbering_o6Pm" style="counter-reset:line-count 0"><span class="token-line codeLine_lJS_" style="color:#393A34"><span class="codeLineNumber_Tfdd"></span><span class="codeLineContent_feaV"><span class="token plain">curl --location 'http://localhost:4000/v1/videos' \</span></span><br></span><span class="token-line codeLine_lJS_" style="color:#393A34"><span class="codeLineNumber_Tfdd"></span><span class="codeLineContent_feaV"><span class="token plain">--header 'Content-Type: application/json' \</span></span><br></span><span class="token-line codeLine_lJS_" style="color:#393A34"><span class="codeLineNumber_Tfdd"></span><span class="codeLineContent_feaV"><span class="token plain">--header 'Authorization: Bearer sk-1234' \</span></span><br></span><span class="token-line codeLine_lJS_" style="color:#393A34"><span class="codeLineNumber_Tfdd"></span><span class="codeLineContent_feaV"><span class="token plain">--data '{</span></span><br></span><span class="token-line codeLine_lJS_" style="color:#393A34"><span class="codeLineNumber_Tfdd"></span><span class="codeLineContent_feaV"><span class="token plain">    "model": "runwayml/gen4_turbo",</span></span><br></span><span class="token-line codeLine_lJS_" style="color:#393A34"><span class="codeLineNumber_Tfdd"></span><span class="codeLineContent_feaV"><span class="token plain">    "prompt": "A high quality demo video of litellm ai gateway",</span></span><br></span><span class="token-line codeLine_lJS_" style="color:#393A34"><span class="codeLineNumber_Tfdd"></span><span class="codeLineContent_feaV"><span class="token plain">    "input_reference": "https://example.com/image.jpg",</span></span><br></span><span class="token-line codeLine_lJS_" style="color:#393A34"><span class="codeLineNumber_Tfdd"></span><span class="codeLineContent_feaV"><span class="token plain">    "seconds": 5,</span></span><br></span><span class="token-line codeLine_lJS_" style="color:#393A34"><span class="codeLineNumber_Tfdd"></span><span class="codeLineContent_feaV"><span class="token plain">    "size": "1280x720"</span></span><br></span><span class="token-line codeLine_lJS_" style="color:#393A34"><span class="codeLineNumber_Tfdd"></span><span class="codeLineContent_feaV"><span class="token plain">}'</span></span><br></span></code></pre></div></div>
<p><a href="https://docs.litellm.ai/docs/providers/runwayml/videos">Get Started with RunwayML</a></p>
<hr>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="prometheus-metrics---open-source">Prometheus Metrics - Open Source<a href="https://docs.litellm.ai/release_notes/v1-80-0#prometheus-metrics---open-source" class="hash-link" aria-label="Direct link to Prometheus Metrics - Open Source" title="Direct link to Prometheus Metrics - Open Source">​</a></h3>
<p>Prometheus metrics are now available in the open-source version of LiteLLM, providing comprehensive observability for your AI Gateway without requiring an enterprise license.</p>
<p><strong>Quick Start:</strong></p>
<div class="language-yaml codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-yaml codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token key atrule" style="color:#00a4db">litellm_settings</span><span class="token punctuation" style="color:#393A34">:</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">  </span><span class="token key atrule" style="color:#00a4db">success_callback</span><span class="token punctuation" style="color:#393A34">:</span><span class="token plain"> </span><span class="token punctuation" style="color:#393A34">[</span><span class="token string" style="color:#e3116c">"prometheus"</span><span class="token punctuation" style="color:#393A34">]</span><span class="token plain"></span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">  </span><span class="token key atrule" style="color:#00a4db">failure_callback</span><span class="token punctuation" style="color:#393A34">:</span><span class="token plain"> </span><span class="token punctuation" style="color:#393A34">[</span><span class="token string" style="color:#e3116c">"prometheus"</span><span class="token punctuation" style="color:#393A34">]</span><br></span></code></pre></div></div>
<p><a href="https://docs.litellm.ai/docs/proxy/logging#prometheus">Get Started with Prometheus</a></p>
<hr>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="vector-store-files-api">Vector Store Files API<a href="https://docs.litellm.ai/release_notes/v1-80-0#vector-store-files-api" class="hash-link" aria-label="Direct link to Vector Store Files API" title="Direct link to Vector Store Files API">​</a></h3>
<p>Complete OpenAI-compatible Vector Store Files API now stable, enabling full file lifecycle management within vector stores.</p>
<p><strong>Supported Endpoints:</strong></p>
<ul>
<li><code>POST /v1/vector_stores/{vector_store_id}/files</code> - Create vector store file</li>
<li><code>GET /v1/vector_stores/{vector_store_id}/files</code> - List vector store files</li>
<li><code>GET /v1/vector_stores/{vector_store_id}/files/{file_id}</code> - Retrieve vector store file</li>
<li><code>GET /v1/vector_stores/{vector_store_id}/files/{file_id}/content</code> - Retrieve file content</li>
<li><code>DELETE /v1/vector_stores/{vector_store_id}/files/{file_id}</code> - Delete vector store file</li>
<li><code>DELETE /v1/vector_stores/{vector_store_id}</code> - Delete vector store</li>
</ul>
<p><strong>Quick Start:</strong></p>
<div class="language-bash codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">Create Vector Store File</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-bash codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv codeBlockLinesWithNumbering_o6Pm" style="counter-reset:line-count 0"><span class="token-line codeLine_lJS_" style="color:#393A34"><span class="codeLineNumber_Tfdd"></span><span class="codeLineContent_feaV"><span class="token plain">curl --location 'http://localhost:4000/v1/vector_stores/vs_123/files' \</span></span><br></span><span class="token-line codeLine_lJS_" style="color:#393A34"><span class="codeLineNumber_Tfdd"></span><span class="codeLineContent_feaV"><span class="token plain">--header 'Content-Type: application/json' \</span></span><br></span><span class="token-line codeLine_lJS_" style="color:#393A34"><span class="codeLineNumber_Tfdd"></span><span class="codeLineContent_feaV"><span class="token plain">--header 'Authorization: Bearer sk-1234' \</span></span><br></span><span class="token-line codeLine_lJS_" style="color:#393A34"><span class="codeLineNumber_Tfdd"></span><span class="codeLineContent_feaV"><span class="token plain">--data '{</span></span><br></span><span class="token-line codeLine_lJS_" style="color:#393A34"><span class="codeLineNumber_Tfdd"></span><span class="codeLineContent_feaV"><span class="token plain">    "file_id": "file_abc"</span></span><br></span><span class="token-line codeLine_lJS_" style="color:#393A34"><span class="codeLineNumber_Tfdd"></span><span class="codeLineContent_feaV"><span class="token plain">}'</span></span><br></span></code></pre></div></div>
<p><a href="https://docs.litellm.ai/docs/vector_store_files">Get Started with Vector Stores</a></p>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-providers-and-endpoints">New Providers and Endpoints<a href="https://docs.litellm.ai/release_notes/v1-80-0#new-providers-and-endpoints" class="hash-link" aria-label="Direct link to New Providers and Endpoints" title="Direct link to New Providers and Endpoints">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="new-providers">New Providers<a href="https://docs.litellm.ai/release_notes/v1-80-0#new-providers" class="hash-link" aria-label="Direct link to New Providers" title="Direct link to New Providers">​</a></h3>
<table><thead><tr><th>Provider</th><th>Supported Endpoints</th><th>Description</th></tr></thead><tbody><tr><td><strong><a href="https://docs.litellm.ai/docs/providers/runwayml/videos">RunwayML</a></strong></td><td><code>/v1/videos</code>, <code>/v1/images/generations</code>, <code>/v1/audio/speech</code></td><td>Gen-4 video generation, image generation, and text-to-speech</td></tr></tbody></table>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="new-llm-api-endpoints">New LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-80-0#new-llm-api-endpoints" class="hash-link" aria-label="Direct link to New LLM API Endpoints" title="Direct link to New LLM API Endpoints">​</a></h3>
<table><thead><tr><th>Endpoint</th><th>Method</th><th>Description</th><th>Documentation</th></tr></thead><tbody><tr><td><code>/v1/vector_stores/{vector_store_id}/files</code></td><td>POST</td><td>Create vector store file</td><td><a href="https://docs.litellm.ai/docs/vector_store_files">Docs</a></td></tr><tr><td><code>/v1/vector_stores/{vector_store_id}/files</code></td><td>GET</td><td>List vector store files</td><td><a href="https://docs.litellm.ai/docs/vector_store_files">Docs</a></td></tr><tr><td><code>/v1/vector_stores/{vector_store_id}/files/{file_id}</code></td><td>GET</td><td>Retrieve vector store file</td><td><a href="https://docs.litellm.ai/docs/vector_store_files">Docs</a></td></tr><tr><td><code>/v1/vector_stores/{vector_store_id}/files/{file_id}/content</code></td><td>GET</td><td>Retrieve file content</td><td><a href="https://docs.litellm.ai/docs/vector_store_files">Docs</a></td></tr><tr><td><code>/v1/vector_stores/{vector_store_id}/files/{file_id}</code></td><td>DELETE</td><td>Delete vector store file</td><td><a href="https://docs.litellm.ai/docs/vector_store_files">Docs</a></td></tr><tr><td><code>/v1/vector_stores/{vector_store_id}</code></td><td>DELETE</td><td>Delete vector store</td><td><a href="https://docs.litellm.ai/docs/vector_store_files">Docs</a></td></tr></tbody></table>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-80-0#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support">New Model Support<a href="https://docs.litellm.ai/release_notes/v1-80-0#new-model-support" class="hash-link" aria-label="Direct link to New Model Support" title="Direct link to New Model Support">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th><th>Features</th></tr></thead><tbody><tr><td>OpenAI</td><td><code>gpt-5.1</code></td><td>272K</td><td>$1.25</td><td>$10.00</td><td>Reasoning, vision, PDF input, responses API</td></tr><tr><td>OpenAI</td><td><code>gpt-5.1-2025-11-13</code></td><td>272K</td><td>$1.25</td><td>$10.00</td><td>Reasoning, vision, PDF input, responses API</td></tr><tr><td>OpenAI</td><td><code>gpt-5.1-chat-latest</code></td><td>128K</td><td>$1.25</td><td>$10.00</td><td>Reasoning, vision, PDF input</td></tr><tr><td>OpenAI</td><td><code>gpt-5.1-codex</code></td><td>272K</td><td>$1.25</td><td>$10.00</td><td>Responses API, reasoning, vision</td></tr><tr><td>OpenAI</td><td><code>gpt-5.1-codex-mini</code></td><td>272K</td><td>$0.25</td><td>$2.00</td><td>Responses API, reasoning, vision</td></tr><tr><td>Moonshot</td><td><code>moonshot/kimi-k2-thinking</code></td><td>262K</td><td>$0.60</td><td>$2.50</td><td>Function calling, web search, reasoning</td></tr><tr><td>Mistral</td><td><code>mistral/magistral-medium-2509</code></td><td>40K</td><td>$2.00</td><td>$5.00</td><td>Reasoning, function calling</td></tr><tr><td>Vertex AI</td><td><code>vertex_ai/moonshotai/kimi-k2-thinking-maas</code></td><td>256K</td><td>$0.60</td><td>$2.50</td><td>Function calling, web search</td></tr><tr><td>OpenRouter</td><td><code>openrouter/deepseek/deepseek-v3.2-exp</code></td><td>164K</td><td>$0.20</td><td>$0.40</td><td>Function calling, prompt caching</td></tr><tr><td>OpenRouter</td><td><code>openrouter/minimax/minimax-m2</code></td><td>205K</td><td>$0.26</td><td>$1.02</td><td>Function calling, reasoning</td></tr><tr><td>OpenRouter</td><td><code>openrouter/z-ai/glm-4.6</code></td><td>203K</td><td>$0.40</td><td>$1.75</td><td>Function calling, reasoning</td></tr><tr><td>OpenRouter</td><td><code>openrouter/z-ai/glm-4.6:exacto</code></td><td>203K</td><td>$0.45</td><td>$1.90</td><td>Function calling, reasoning</td></tr><tr><td>Voyage</td><td><code>voyage/voyage-3.5</code></td><td>32K</td><td>$0.06</td><td>-</td><td>Embeddings</td></tr><tr><td>Voyage</td><td><code>voyage/voyage-3.5-lite</code></td><td>32K</td><td>$0.02</td><td>-</td><td>Embeddings</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="video-generation-models">Video Generation Models<a href="https://docs.litellm.ai/release_notes/v1-80-0#video-generation-models" class="hash-link" aria-label="Direct link to Video Generation Models" title="Direct link to Video Generation Models">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Cost Per Second</th><th>Resolutions</th><th>Features</th></tr></thead><tbody><tr><td>RunwayML</td><td><code>runwayml/gen4_turbo</code></td><td>$0.05</td><td>1280x720, 720x1280</td><td>Text + image to video</td></tr><tr><td>RunwayML</td><td><code>runwayml/gen4_aleph</code></td><td>$0.15</td><td>1280x720, 720x1280</td><td>Text + image to video</td></tr><tr><td>RunwayML</td><td><code>runwayml/gen3a_turbo</code></td><td>$0.05</td><td>1280x720, 720x1280</td><td>Text + image to video</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="image-generation-models">Image Generation Models<a href="https://docs.litellm.ai/release_notes/v1-80-0#image-generation-models" class="hash-link" aria-label="Direct link to Image Generation Models" title="Direct link to Image Generation Models">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Cost Per Image</th><th>Resolutions</th><th>Features</th></tr></thead><tbody><tr><td>RunwayML</td><td><code>runwayml/gen4_image</code></td><td>$0.05</td><td>1280x720, 1920x1080</td><td>Text + image to image</td></tr><tr><td>RunwayML</td><td><code>runwayml/gen4_image_turbo</code></td><td>$0.02</td><td>1280x720, 1920x1080</td><td>Text + image to image</td></tr><tr><td>Fal.ai</td><td><code>fal_ai/fal-ai/flux-pro/v1.1</code></td><td>$0.04/image</td><td>-</td><td>Image generation</td></tr><tr><td>Fal.ai</td><td><code>fal_ai/fal-ai/flux/schnell</code></td><td>$0.003/image</td><td>-</td><td>Fast image generation</td></tr><tr><td>Fal.ai</td><td><code>fal_ai/fal-ai/bytedance/seedream/v3/text-to-image</code></td><td>$0.03/image</td><td>-</td><td>Image generation</td></tr><tr><td>Fal.ai</td><td><code>fal_ai/fal-ai/bytedance/dreamina/v3.1/text-to-image</code></td><td>$0.03/image</td><td>-</td><td>Image generation</td></tr><tr><td>Fal.ai</td><td><code>fal_ai/fal-ai/ideogram/v3</code></td><td>$0.06/image</td><td>-</td><td>Image generation</td></tr><tr><td>Fal.ai</td><td><code>fal_ai/fal-ai/imagen4/preview/fast</code></td><td>$0.02/image</td><td>-</td><td>Fast image generation</td></tr><tr><td>Fal.ai</td><td><code>fal_ai/fal-ai/imagen4/preview/ultra</code></td><td>$0.06/image</td><td>-</td><td>High-quality image generation</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="audio-models">Audio Models<a href="https://docs.litellm.ai/release_notes/v1-80-0#audio-models" class="hash-link" aria-label="Direct link to Audio Models" title="Direct link to Audio Models">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Cost</th><th>Features</th></tr></thead><tbody><tr><td>RunwayML</td><td><code>runwayml/eleven_multilingual_v2</code></td><td>$0.0003/char</td><td>Text-to-speech</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-80-0#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI</a></strong></p>
<ul>
<li>Add GPT-5.1 family support with reasoning capabilities - <a href="https://github.com/BerriAI/litellm/pull/16598" target="_blank" rel="noopener noreferrer">PR #16598</a></li>
<li>Add support for <code>reasoning_effort='none'</code> for GPT-5.1 - <a href="https://github.com/BerriAI/litellm/pull/16658" target="_blank" rel="noopener noreferrer">PR #16658</a></li>
<li>Add <code>verbosity</code> parameter support for GPT-5 family models - <a href="https://github.com/BerriAI/litellm/pull/16660" target="_blank" rel="noopener noreferrer">PR #16660</a></li>
<li>Fix forward OpenAI organization for image generation - <a href="https://github.com/BerriAI/litellm/pull/16607" target="_blank" rel="noopener noreferrer">PR #16607</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Gemini (Google AI Studio + Vertex AI)</a></strong></p>
<ul>
<li>Add support for <code>reasoning_effort='none'</code> for Gemini models - <a href="https://github.com/BerriAI/litellm/pull/16548" target="_blank" rel="noopener noreferrer">PR #16548</a></li>
<li>Add all Gemini image models support in image generation - <a href="https://github.com/BerriAI/litellm/pull/16526" target="_blank" rel="noopener noreferrer">PR #16526</a></li>
<li>Add Gemini image edit support - <a href="https://github.com/BerriAI/litellm/pull/16430" target="_blank" rel="noopener noreferrer">PR #16430</a></li>
<li>Fix preserve non-ASCII characters in function call arguments - <a href="https://github.com/BerriAI/litellm/pull/16550" target="_blank" rel="noopener noreferrer">PR #16550</a></li>
<li>Fix Gemini conversation format issue with MCP auto-execution - <a href="https://github.com/BerriAI/litellm/pull/16592" target="_blank" rel="noopener noreferrer">PR #16592</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong></p>
<ul>
<li>Add support for filtering knowledge base queries - <a href="https://github.com/BerriAI/litellm/pull/16543" target="_blank" rel="noopener noreferrer">PR #16543</a></li>
<li>Ensure correct <code>aws_region</code> is used when provided dynamically for embeddings - <a href="https://github.com/BerriAI/litellm/pull/16547" target="_blank" rel="noopener noreferrer">PR #16547</a></li>
<li>Add support for custom KMS encryption keys in Bedrock Batch operations - <a href="https://github.com/BerriAI/litellm/pull/16662" target="_blank" rel="noopener noreferrer">PR #16662</a></li>
<li>Add bearer token authentication support for AgentCore - <a href="https://github.com/BerriAI/litellm/pull/16556" target="_blank" rel="noopener noreferrer">PR #16556</a></li>
<li>Fix AgentCore SSE stream iterator to async for proper streaming support - <a href="https://github.com/BerriAI/litellm/pull/16293" target="_blank" rel="noopener noreferrer">PR #16293</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Add context management param support - <a href="https://github.com/BerriAI/litellm/pull/16528" target="_blank" rel="noopener noreferrer">PR #16528</a></li>
<li>Fix preserve <code>$defs</code> for Anthropic tools input schema - <a href="https://github.com/BerriAI/litellm/pull/16648" target="_blank" rel="noopener noreferrer">PR #16648</a></li>
<li>Fix support Anthropic tool_use and tool_result in token counter - <a href="https://github.com/BerriAI/litellm/pull/16351" target="_blank" rel="noopener noreferrer">PR #16351</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vertex_ai">Vertex AI</a></strong></p>
<ul>
<li>Add Vertex Kimi-K2-Thinking support - <a href="https://github.com/BerriAI/litellm/pull/16671" target="_blank" rel="noopener noreferrer">PR #16671</a></li>
<li>Add <code>vertex_credentials</code> support to <code>litellm.rerank()</code> - <a href="https://github.com/BerriAI/litellm/pull/16479" target="_blank" rel="noopener noreferrer">PR #16479</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/mistral">Mistral</a></strong></p>
<ul>
<li>Fix Magistral streaming to emit reasoning chunks - <a href="https://github.com/BerriAI/litellm/pull/16434" target="_blank" rel="noopener noreferrer">PR #16434</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/moonshot">Moonshot (Kimi)</a></strong></p>
<ul>
<li>Add Kimi K2 thinking model support - <a href="https://github.com/BerriAI/litellm/pull/16445" target="_blank" rel="noopener noreferrer">PR #16445</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/sambanova">SambaNova</a></strong></p>
<ul>
<li>Fix SambaNova API rejecting requests when message content is passed as a list format - <a href="https://github.com/BerriAI/litellm/pull/16612" target="_blank" rel="noopener noreferrer">PR #16612</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vllm">VLLM</a></strong></p>
<ul>
<li>Fix use vllm passthrough config for hosted vllm provider instead of raising error - <a href="https://github.com/BerriAI/litellm/pull/16537" target="_blank" rel="noopener noreferrer">PR #16537</a></li>
<li>Add headers to VLLM Passthrough requests with success event logging - <a href="https://github.com/BerriAI/litellm/pull/16532" target="_blank" rel="noopener noreferrer">PR #16532</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/azure">Azure</a></strong></p>
<ul>
<li>Fix improve Azure auth parameter handling for None values - <a href="https://github.com/BerriAI/litellm/pull/14436" target="_blank" rel="noopener noreferrer">PR #14436</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/groq">Groq</a></strong></p>
<ul>
<li>Fix parse failed chunks for Groq - <a href="https://github.com/BerriAI/litellm/pull/16595" target="_blank" rel="noopener noreferrer">PR #16595</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/voyage">Voyage</a></strong></p>
<ul>
<li>Add Voyage 3.5 and 3.5-lite embeddings pricing and doc update - <a href="https://github.com/BerriAI/litellm/pull/16641" target="_blank" rel="noopener noreferrer">PR #16641</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/image_generation">Fal.ai</a></strong></p>
<ul>
<li>Add fal-ai/flux/schnell support - <a href="https://github.com/BerriAI/litellm/pull/16580" target="_blank" rel="noopener noreferrer">PR #16580</a></li>
<li>Add all Imagen4 variants of fal ai in model map - <a href="https://github.com/BerriAI/litellm/pull/16579" target="_blank" rel="noopener noreferrer">PR #16579</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-80-0#bug-fixes" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li><strong>General</strong>
<ul>
<li>Fix sanitize null token usage in OpenAI-compatible responses - <a href="https://github.com/BerriAI/litellm/pull/16493" target="_blank" rel="noopener noreferrer">PR #16493</a></li>
<li>Fix apply provided timeout value to ClientTimeout.total - <a href="https://github.com/BerriAI/litellm/pull/16395" target="_blank" rel="noopener noreferrer">PR #16395</a></li>
<li>Fix raising wrong 429 error on wrong exception - <a href="https://github.com/BerriAI/litellm/pull/16482" target="_blank" rel="noopener noreferrer">PR #16482</a></li>
<li>Add new models, delete repeat models, update pricing - <a href="https://github.com/BerriAI/litellm/pull/16491" target="_blank" rel="noopener noreferrer">PR #16491</a></li>
<li>Update model logging format for custom LLM provider - <a href="https://github.com/BerriAI/litellm/pull/16485" target="_blank" rel="noopener noreferrer">PR #16485</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-80-0#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-endpoints">New Endpoints<a href="https://docs.litellm.ai/release_notes/v1-80-0#new-endpoints" class="hash-link" aria-label="Direct link to New Endpoints" title="Direct link to New Endpoints">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/management_endpoints">GET /providers</a></strong>
<ul>
<li>Add GET list of providers endpoint - <a href="https://github.com/BerriAI/litellm/pull/16432" target="_blank" rel="noopener noreferrer">PR #16432</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-80-0#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/video_generation">Video Generation API</a></strong></p>
<ul>
<li>Allow internal users to access video generation routes - <a href="https://github.com/BerriAI/litellm/pull/16472" target="_blank" rel="noopener noreferrer">PR #16472</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/vector_stores">Vector Stores API</a></strong></p>
<ul>
<li>Vector store files stable release with complete CRUD operations - <a href="https://github.com/BerriAI/litellm/pull/16643" target="_blank" rel="noopener noreferrer">PR #16643</a>
<ul>
<li><code>POST /v1/vector_stores/{vector_store_id}/files</code> - Create vector store file</li>
<li><code>GET /v1/vector_stores/{vector_store_id}/files</code> - List vector store files</li>
<li><code>GET /v1/vector_stores/{vector_store_id}/files/{file_id}</code> - Retrieve vector store file</li>
<li><code>GET /v1/vector_stores/{vector_store_id}/files/{file_id}/content</code> - Retrieve file content</li>
<li><code>DELETE /v1/vector_stores/{vector_store_id}/files/{file_id}</code> - Delete vector store file</li>
<li><code>DELETE /v1/vector_stores/{vector_store_id}</code> - Delete vector store</li>
</ul>
</li>
<li>Ensure users can access <code>search_results</code> for both stream + non-stream response - <a href="https://github.com/BerriAI/litellm/pull/16459" target="_blank" rel="noopener noreferrer">PR #16459</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs">Bugs<a href="https://docs.litellm.ai/release_notes/v1-80-0#bugs" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/video_generation">Video Generation API</a></strong></p>
<ul>
<li>Fix use GET for <code>/v1/videos/{video_id}/content</code> - <a href="https://github.com/BerriAI/litellm/pull/16672" target="_blank" rel="noopener noreferrer">PR #16672</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Fix remove generic exception handling - <a href="https://github.com/BerriAI/litellm/pull/16599" target="_blank" rel="noopener noreferrer">PR #16599</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-80-0#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-80-0#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong>Proxy CLI Auth</strong></p>
<ul>
<li>Fix remove strict master_key check in add_deployment - <a href="https://github.com/BerriAI/litellm/pull/16453" target="_blank" rel="noopener noreferrer">PR #16453</a></li>
</ul>
</li>
<li>
<p><strong>Virtual Keys</strong></p>
<ul>
<li>UI - Add Tags To Edit Key Flow - <a href="https://github.com/BerriAI/litellm/pull/16500" target="_blank" rel="noopener noreferrer">PR #16500</a></li>
<li>UI - Test Key Page show models based on selected endpoint - <a href="https://github.com/BerriAI/litellm/pull/16452" target="_blank" rel="noopener noreferrer">PR #16452</a></li>
<li>UI - Expose user_alias in view and update path - <a href="https://github.com/BerriAI/litellm/pull/16669" target="_blank" rel="noopener noreferrer">PR #16669</a></li>
</ul>
</li>
<li>
<p><strong>Models + Endpoints</strong></p>
<ul>
<li>UI - Add LiteLLM Params to Edit Model - <a href="https://github.com/BerriAI/litellm/pull/16496" target="_blank" rel="noopener noreferrer">PR #16496</a></li>
<li>UI - Add Model use backend data - <a href="https://github.com/BerriAI/litellm/pull/16664" target="_blank" rel="noopener noreferrer">PR #16664</a></li>
<li>UI - Remove Description Field from LLM Credentials - <a href="https://github.com/BerriAI/litellm/pull/16608" target="_blank" rel="noopener noreferrer">PR #16608</a></li>
<li>UI - Add RunwayML on Admin UI supported models/providers - <a href="https://github.com/BerriAI/litellm/pull/16606" target="_blank" rel="noopener noreferrer">PR #16606</a></li>
<li>Infra - Migrate Add Model Fields to Backend - <a href="https://github.com/BerriAI/litellm/pull/16620" target="_blank" rel="noopener noreferrer">PR #16620</a></li>
<li>Add API Endpoint for creating model access group - <a href="https://github.com/BerriAI/litellm/pull/16663" target="_blank" rel="noopener noreferrer">PR #16663</a></li>
</ul>
</li>
<li>
<p><strong>Teams</strong></p>
<ul>
<li>UI - Invite User Searchable Team Select - <a href="https://github.com/BerriAI/litellm/pull/16454" target="_blank" rel="noopener noreferrer">PR #16454</a></li>
<li>Fix use user budget instead of key budget when creating new team - <a href="https://github.com/BerriAI/litellm/pull/16074" target="_blank" rel="noopener noreferrer">PR #16074</a></li>
</ul>
</li>
<li>
<p><strong>Budgets</strong></p>
<ul>
<li>UI - Move Budgets out of Experimental - <a href="https://github.com/BerriAI/litellm/pull/16544" target="_blank" rel="noopener noreferrer">PR #16544</a></li>
</ul>
</li>
<li>
<p><strong>Guardrails</strong></p>
<ul>
<li>UI - Config Guardrails should not be deletable from table - <a href="https://github.com/BerriAI/litellm/pull/16540" target="_blank" rel="noopener noreferrer">PR #16540</a></li>
<li>Fix remove enterprise restriction from guardrails list endpoint - <a href="https://github.com/BerriAI/litellm/pull/15333" target="_blank" rel="noopener noreferrer">PR #15333</a></li>
</ul>
</li>
<li>
<p><strong>Callbacks</strong></p>
<ul>
<li>UI - New Callbacks table - <a href="https://github.com/BerriAI/litellm/pull/16512" target="_blank" rel="noopener noreferrer">PR #16512</a></li>
<li>Fix delete callbacks failing - <a href="https://github.com/BerriAI/litellm/pull/16473" target="_blank" rel="noopener noreferrer">PR #16473</a></li>
</ul>
</li>
<li>
<p><strong>Usage &amp; Analytics</strong></p>
<ul>
<li>UI - Improve Usage Indicator - <a href="https://github.com/BerriAI/litellm/pull/16504" target="_blank" rel="noopener noreferrer">PR #16504</a></li>
<li>UI - Model Info Page Health Check - <a href="https://github.com/BerriAI/litellm/pull/16416" target="_blank" rel="noopener noreferrer">PR #16416</a></li>
<li>Infra - Show Deprecation Warning for Model Analytics Tab - <a href="https://github.com/BerriAI/litellm/pull/16417" target="_blank" rel="noopener noreferrer">PR #16417</a></li>
<li>Fix Litellm tags usage add request_id - <a href="https://github.com/BerriAI/litellm/pull/16111" target="_blank" rel="noopener noreferrer">PR #16111</a></li>
</ul>
</li>
<li>
<p><strong>Health Check</strong></p>
<ul>
<li>Add Langfuse OTEL and SQS to Health Check - <a href="https://github.com/BerriAI/litellm/pull/16514" target="_blank" rel="noopener noreferrer">PR #16514</a></li>
</ul>
</li>
<li>
<p><strong>General UI</strong></p>
<ul>
<li>UI - Normalize table action columns appearance - <a href="https://github.com/BerriAI/litellm/pull/16657" target="_blank" rel="noopener noreferrer">PR #16657</a></li>
<li>UI - Button Styles and Sizing in Settings Pages - <a href="https://github.com/BerriAI/litellm/pull/16600" target="_blank" rel="noopener noreferrer">PR #16600</a></li>
<li>UI - SSO Modal Cosmetic Changes - <a href="https://github.com/BerriAI/litellm/pull/16554" target="_blank" rel="noopener noreferrer">PR #16554</a></li>
<li>Fix UI logos loading with SERVER_ROOT_PATH - <a href="https://github.com/BerriAI/litellm/pull/16618" target="_blank" rel="noopener noreferrer">PR #16618</a></li>
<li>Fix remove misleading 'Custom' option mention from OpenAI endpoint tooltips - <a href="https://github.com/BerriAI/litellm/pull/16622" target="_blank" rel="noopener noreferrer">PR #16622</a></li>
</ul>
</li>
<li>
<p><strong>SSO</strong></p>
<ul>
<li>Ensure <code>role</code> from SSO provider is used when a user is inserted onto LiteLLM - <a href="https://github.com/BerriAI/litellm/pull/16794" target="_blank" rel="noopener noreferrer">PR #16794</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs-1">Bugs<a href="https://docs.litellm.ai/release_notes/v1-80-0#bugs-1" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>Management Endpoints</strong>
<ul>
<li>Fix inconsistent error responses in customer management endpoints - <a href="https://github.com/BerriAI/litellm/pull/16450" target="_blank" rel="noopener noreferrer">PR #16450</a></li>
<li>Fix correct date range filtering in /spend/logs endpoint - <a href="https://github.com/BerriAI/litellm/pull/16443" target="_blank" rel="noopener noreferrer">PR #16443</a></li>
<li>Fix /spend/logs/ui Access Control - <a href="https://github.com/BerriAI/litellm/pull/16446" target="_blank" rel="noopener noreferrer">PR #16446</a></li>
<li>Add pagination for /spend/logs/session/ui endpoint - <a href="https://github.com/BerriAI/litellm/pull/16603" target="_blank" rel="noopener noreferrer">PR #16603</a></li>
<li>Fix LiteLLM Usage shows key_hash - <a href="https://github.com/BerriAI/litellm/pull/16471" target="_blank" rel="noopener noreferrer">PR #16471</a></li>
<li>Fix app_roles missing from jwt payload - <a href="https://github.com/BerriAI/litellm/pull/16448" target="_blank" rel="noopener noreferrer">PR #16448</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="logging--guardrail--prompt-management-integrations">Logging / Guardrail / Prompt Management Integrations<a href="https://docs.litellm.ai/release_notes/v1-80-0#logging--guardrail--prompt-management-integrations" class="hash-link" aria-label="Direct link to Logging / Guardrail / Prompt Management Integrations" title="Direct link to Logging / Guardrail / Prompt Management Integrations">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-integration">New Integration<a href="https://docs.litellm.ai/release_notes/v1-80-0#new-integration" class="hash-link" aria-label="Direct link to New Integration" title="Direct link to New Integration">​</a></h4>
<ul>
<li><strong>🆕 <a href="https://docs.litellm.ai/docs/proxy/guardrails/zscaler_ai_guard">Zscaler AI Guard</a></strong>
<ul>
<li>Add Zscaler AI Guard hook for security policy enforcement - <a href="https://github.com/BerriAI/litellm/pull/15691" target="_blank" rel="noopener noreferrer">PR #15691</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="logging">Logging<a href="https://docs.litellm.ai/release_notes/v1-80-0#logging" class="hash-link" aria-label="Direct link to Logging" title="Direct link to Logging">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#langfuse">Langfuse</a></strong></p>
<ul>
<li>Fix handle null usage values to prevent validation errors - <a href="https://github.com/BerriAI/litellm/pull/16396" target="_blank" rel="noopener noreferrer">PR #16396</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging">CloudZero</a></strong></p>
<ul>
<li>Fix updated spend would not be sent to CloudZero - <a href="https://github.com/BerriAI/litellm/pull/16201" target="_blank" rel="noopener noreferrer">PR #16201</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails">Guardrails<a href="https://docs.litellm.ai/release_notes/v1-80-0#guardrails" class="hash-link" aria-label="Direct link to Guardrails" title="Direct link to Guardrails">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">IBM Detector</a></strong>
<ul>
<li>Ensure detector-id is passed as header to IBM detector server - <a href="https://github.com/BerriAI/litellm/pull/16649" target="_blank" rel="noopener noreferrer">PR #16649</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="prompt-management">Prompt Management<a href="https://docs.litellm.ai/release_notes/v1-80-0#prompt-management" class="hash-link" aria-label="Direct link to Prompt Management" title="Direct link to Prompt Management">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/prompt_management">Custom Prompt Management</a></strong>
<ul>
<li>Add SDK focused examples for custom prompt management - <a href="https://github.com/BerriAI/litellm/pull/16441" target="_blank" rel="noopener noreferrer">PR #16441</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-80-0#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li><strong>End User Budgets</strong>
<ul>
<li>Allow pointing max_end_user budget to an id, so the default ID applies to all end users - <a href="https://github.com/BerriAI/litellm/pull/16456" target="_blank" rel="noopener noreferrer">PR #16456</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-gateway">MCP Gateway<a href="https://docs.litellm.ai/release_notes/v1-80-0#mcp-gateway" class="hash-link" aria-label="Direct link to MCP Gateway" title="Direct link to MCP Gateway">​</a></h2>
<ul>
<li><strong>Configuration</strong>
<ul>
<li>Add dynamic OAuth2 metadata discovery for MCP servers - <a href="https://github.com/BerriAI/litellm/pull/16676" target="_blank" rel="noopener noreferrer">PR #16676</a></li>
<li>Fix allow tool call even when server name prefix is missing - <a href="https://github.com/BerriAI/litellm/pull/16425" target="_blank" rel="noopener noreferrer">PR #16425</a></li>
<li>Fix exclude unauthorized MCP servers from allowed server list - <a href="https://github.com/BerriAI/litellm/pull/16551" target="_blank" rel="noopener noreferrer">PR #16551</a></li>
<li>Fix unable to delete MCP server from permission settings - <a href="https://github.com/BerriAI/litellm/pull/16407" target="_blank" rel="noopener noreferrer">PR #16407</a></li>
<li>Fix avoid crashing when MCP server record lacks credentials - <a href="https://github.com/BerriAI/litellm/pull/16601" target="_blank" rel="noopener noreferrer">PR #16601</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="agents">Agents<a href="https://docs.litellm.ai/release_notes/v1-80-0#agents" class="hash-link" aria-label="Direct link to Agents" title="Direct link to Agents">​</a></h2>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/agents">Agent Registration (A2A Spec)</a></strong>
<ul>
<li>Support agent registration + discovery following Agent-to-Agent specification - <a href="https://github.com/BerriAI/litellm/pull/16615" target="_blank" rel="noopener noreferrer">PR #16615</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements">Performance / Loadbalancing / Reliability improvements<a href="https://docs.litellm.ai/release_notes/v1-80-0#performance--loadbalancing--reliability-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements" title="Direct link to Performance / Loadbalancing / Reliability improvements">​</a></h2>
<ul>
<li>
<p><strong>Embeddings Performance</strong></p>
<ul>
<li>Use router's O(1) lookup and shared sessions for embeddings - <a href="https://github.com/BerriAI/litellm/pull/16344" target="_blank" rel="noopener noreferrer">PR #16344</a></li>
</ul>
</li>
<li>
<p><strong>Router Reliability</strong></p>
<ul>
<li>Support default fallbacks for unknown models - <a href="https://github.com/BerriAI/litellm/pull/16419" target="_blank" rel="noopener noreferrer">PR #16419</a></li>
</ul>
</li>
<li>
<p><strong>Callback Management</strong></p>
<ul>
<li>Add atexit handlers to flush callbacks for async completions - <a href="https://github.com/BerriAI/litellm/pull/16487" target="_blank" rel="noopener noreferrer">PR #16487</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="general-proxy-improvements">General Proxy Improvements<a href="https://docs.litellm.ai/release_notes/v1-80-0#general-proxy-improvements" class="hash-link" aria-label="Direct link to General Proxy Improvements" title="Direct link to General Proxy Improvements">​</a></h2>
<ul>
<li><strong>Configuration Management</strong>
<ul>
<li>Fix update model_cost_map_url to use environment variable - <a href="https://github.com/BerriAI/litellm/pull/16429" target="_blank" rel="noopener noreferrer">PR #16429</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates">Documentation Updates<a href="https://docs.litellm.ai/release_notes/v1-80-0#documentation-updates" class="hash-link" aria-label="Direct link to Documentation Updates" title="Direct link to Documentation Updates">​</a></h2>
<ul>
<li>
<p><strong>Provider Documentation</strong></p>
<ul>
<li>Fix streaming example in README - <a href="https://github.com/BerriAI/litellm/pull/16461" target="_blank" rel="noopener noreferrer">PR #16461</a></li>
<li>Update broken Slack invite links to support page - <a href="https://github.com/BerriAI/litellm/pull/16546" target="_blank" rel="noopener noreferrer">PR #16546</a></li>
<li>Fix code block indentation for fallbacks page - <a href="https://github.com/BerriAI/litellm/pull/16542" target="_blank" rel="noopener noreferrer">PR #16542</a></li>
<li>Documentation code example corrections - <a href="https://github.com/BerriAI/litellm/pull/16502" target="_blank" rel="noopener noreferrer">PR #16502</a></li>
<li>Document <code>reasoning_effort</code> summary field options - <a href="https://github.com/BerriAI/litellm/pull/16549" target="_blank" rel="noopener noreferrer">PR #16549</a></li>
</ul>
</li>
<li>
<p><strong>API Documentation</strong></p>
<ul>
<li>Add docs on APIs for model access management - <a href="https://github.com/BerriAI/litellm/pull/16673" target="_blank" rel="noopener noreferrer">PR #16673</a></li>
<li>Add docs for showing how to auto reload new pricing data - <a href="https://github.com/BerriAI/litellm/pull/16675" target="_blank" rel="noopener noreferrer">PR #16675</a></li>
<li>LiteLLM Quick start - show how model resolution works - <a href="https://github.com/BerriAI/litellm/pull/16602" target="_blank" rel="noopener noreferrer">PR #16602</a></li>
<li>Add docs for tracking callback failure - <a href="https://github.com/BerriAI/litellm/pull/16474" target="_blank" rel="noopener noreferrer">PR #16474</a></li>
</ul>
</li>
<li>
<p><strong>General Documentation</strong></p>
<ul>
<li>Fix container api link in release page - <a href="https://github.com/BerriAI/litellm/pull/16440" target="_blank" rel="noopener noreferrer">PR #16440</a></li>
<li>Add softgen to projects that are using litellm - <a href="https://github.com/BerriAI/litellm/pull/16423" target="_blank" rel="noopener noreferrer">PR #16423</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-80-0#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@artplan1 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16423" target="_blank" rel="noopener noreferrer">PR #16423</a></li>
<li>@JehandadK made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16472" target="_blank" rel="noopener noreferrer">PR #16472</a></li>
<li>@vmiscenko made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16453" target="_blank" rel="noopener noreferrer">PR #16453</a></li>
<li>@mcowger made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16429" target="_blank" rel="noopener noreferrer">PR #16429</a></li>
<li>@yellowsubmarine372 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16395" target="_blank" rel="noopener noreferrer">PR #16395</a></li>
<li>@Hebruwu made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16201" target="_blank" rel="noopener noreferrer">PR #16201</a></li>
<li>@jwang-gif made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15691" target="_blank" rel="noopener noreferrer">PR #15691</a></li>
<li>@AnthonyMonaco made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16502" target="_blank" rel="noopener noreferrer">PR #16502</a></li>
<li>@andrewm4894 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16487" target="_blank" rel="noopener noreferrer">PR #16487</a></li>
<li>@f14-bertolotti made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16485" target="_blank" rel="noopener noreferrer">PR #16485</a></li>
<li>@busla made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16293" target="_blank" rel="noopener noreferrer">PR #16293</a></li>
<li>@MightyGoldenOctopus made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16537" target="_blank" rel="noopener noreferrer">PR #16537</a></li>
<li>@ultmaster made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14436" target="_blank" rel="noopener noreferrer">PR #14436</a></li>
<li>@bchrobot made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16542" target="_blank" rel="noopener noreferrer">PR #16542</a></li>
<li>@sep-grindr made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16622" target="_blank" rel="noopener noreferrer">PR #16622</a></li>
<li>@pnookala-godaddy made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16607" target="_blank" rel="noopener noreferrer">PR #16607</a></li>
<li>@dtunikov made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16592" target="_blank" rel="noopener noreferrer">PR #16592</a></li>
<li>@lukapecnik made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16648" target="_blank" rel="noopener noreferrer">PR #16648</a></li>
<li>@jyeros made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16618" target="_blank" rel="noopener noreferrer">PR #16618</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog">Full Changelog<a href="https://docs.litellm.ai/release_notes/v1-80-0#full-changelog" class="hash-link" aria-label="Direct link to Full Changelog" title="Direct link to Full Changelog">​</a></h2>
<p><strong><a href="https://github.com/BerriAI/litellm/compare/v1.79.3.rc.1...v1.80.0.rc.1" target="_blank" rel="noopener noreferrer">View complete changelog on GitHub</a></strong></p>
<hr>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[v1.79.3-stable - Built-in Guardrails on AI Gateway]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-79-3</id>
        <link href="https://docs.litellm.ai/release_notes/v1-79-3"/>
        <updated>2025-11-08T10:00:00.000Z</updated>
        <summary type="html"><![CDATA[Deploy this version]]></summary>
        <content type="html"><![CDATA[<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-79-3#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">docker.litellm.ai/berriai/litellm:v1.79.3-stable</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.79.3.rc.1</span><br></span></code></pre></div></div></div></div></div>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-79-3#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<ul>
<li><strong>LiteLLM Custom Guardrail</strong> - Built-in guardrail with UI configuration support</li>
<li><strong>Performance Improvements</strong> - <code>/responses</code> API 19× Lower Median Latency</li>
<li><strong>Veo3 Video Generation (Vertex AI + Google AI Studio)</strong> - Use OpenAI Video API to generate videos with Vertex AI and Google AI Studio Veo3 models</li>
</ul>
<hr>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="built-in-guardrails-on-ai-gateway">Built-in Guardrails on AI Gateway<a href="https://docs.litellm.ai/release_notes/v1-79-3#built-in-guardrails-on-ai-gateway" class="hash-link" aria-label="Direct link to Built-in Guardrails on AI Gateway" title="Direct link to Built-in Guardrails on AI Gateway">​</a></h3>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAFCAYAAAB8ZH1oAAAACXBIWXMAACxLAAAsSwGlPZapAAAAhklEQVR4nDWOQRKDIBAE+f+/kmd4CUGtkoMRwu4CdgpTHvo20zNuWRZCCMzzTEqJIoVWhaKNIkqtFTPDxRgZYe89OWfW9UN472yfzn58MVNEFdda40/nPBvTJDyeB3HL5CyXrfeOU1VuqhnJKlnyZRqzpRTOarjRuBlBa+NTvRaGSaThX8YP/7bBZ2wCQWwAAAAASUVORK5CYII=&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="334"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/built_in_guard.2572eb5.640.png" srcset="/assets/ideal-img/built_in_guard.2572eb5.640.png 640w,/assets/ideal-img/built_in_guard.ba03d25.1920.png 1920w" width="640" height="334"></noscript></div>
<br>
<p>This release introduces built-in guardrails for LiteLLM AI Gateway, allowing you to enforce protections without depending on an external guardrail API.</p>
<ul>
<li><strong>Blocking Keywords</strong> - Block known sensitive keywords like "litellm", "python", etc.</li>
<li><strong>Pattern Detection</strong> - Block known sensitive patterns like emails, Social Security Numbers, API keys, etc.</li>
<li><strong>Custom Regex Patterns</strong> - Define custom regex patterns for your specific use case.</li>
</ul>
<p>Get started with the built-in guardrails on AI Gateway <a href="https://docs.litellm.ai/docs/proxy/guardrails/litellm_content_filter" target="_blank" rel="noopener noreferrer">here</a>.</p>
<hr>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="performance--responses-19-lower-median-latency">Performance – <code>/responses</code> 19× Lower Median Latency<a href="https://docs.litellm.ai/release_notes/v1-79-3#performance--responses-19-lower-median-latency" class="hash-link" aria-label="Direct link to performance--responses-19-lower-median-latency" title="Direct link to performance--responses-19-lower-median-latency">​</a></h3>
<p>This update significantly improves <code>/responses</code> latency by integrating our internal network management for connection handling, eliminating per-request setup overhead.</p>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="results">Results<a href="https://docs.litellm.ai/release_notes/v1-79-3#results" class="hash-link" aria-label="Direct link to Results" title="Direct link to Results">​</a></h4>
<table><thead><tr><th>Metric</th><th>Before</th><th>After</th><th>Improvement</th></tr></thead><tbody><tr><td>Median latency</td><td>3,600 ms</td><td><strong>190 ms</strong></td><td><strong>−95% (~19× faster)</strong></td></tr><tr><td>p95 latency</td><td>4,300 ms</td><td><strong>280 ms</strong></td><td>−93%</td></tr><tr><td>p99 latency</td><td>4,600 ms</td><td><strong>590 ms</strong></td><td>−87%</td></tr><tr><td>Average latency</td><td>3,571 ms</td><td><strong>208 ms</strong></td><td>−94%</td></tr><tr><td>RPS</td><td>231</td><td><strong>1,059</strong></td><td>+358%</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="test-setup">Test Setup<a href="https://docs.litellm.ai/release_notes/v1-79-3#test-setup" class="hash-link" aria-label="Direct link to Test Setup" title="Direct link to Test Setup">​</a></h4>
<table><thead><tr><th>Category</th><th>Specification</th></tr></thead><tbody><tr><td><strong>Load Testing</strong></td><td>Locust: 1,000 concurrent users, 500 ramp-up</td></tr><tr><td><strong>System</strong></td><td>4 vCPUs, 8 GB RAM, 4 workers, 4 instances</td></tr><tr><td><strong>Database</strong></td><td>PostgreSQL (Redis unused)</td></tr><tr><td><strong>Configuration</strong></td><td><a href="https://gist.github.com/AlexsanderHamir/550791675fd752befcac6a9e44024652" target="_blank" rel="noopener noreferrer">config.yaml</a></td></tr><tr><td><strong>Load Script</strong></td><td><a href="https://gist.github.com/AlexsanderHamir/99d673bf74cdd81fd39f59fa9048f2e8" target="_blank" rel="noopener noreferrer">no_cache_hits.py</a></td></tr></tbody></table>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-79-3#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support">New Model Support<a href="https://docs.litellm.ai/release_notes/v1-79-3#new-model-support" class="hash-link" aria-label="Direct link to New Model Support" title="Direct link to New Model Support">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th><th>Features</th></tr></thead><tbody><tr><td>Azure</td><td><code>azure/gpt-5-pro</code></td><td>272K</td><td>$15.00</td><td>$120.00</td><td>Responses API, reasoning, vision, PDF input</td></tr><tr><td>Azure</td><td><code>azure/gpt-image-1-mini</code></td><td>-</td><td>-</td><td>-</td><td>Image generation - per pixel pricing</td></tr><tr><td>Azure</td><td><code>azure/container</code></td><td>-</td><td>-</td><td>-</td><td>Container API - $0.03/session</td></tr><tr><td>OpenAI</td><td><code>openai/container</code></td><td>-</td><td>-</td><td>-</td><td>Container API - $0.03/session</td></tr><tr><td>Cohere</td><td><code>cohere/embed-v4.0</code></td><td>128K</td><td>$0.12</td><td>-</td><td>Embeddings with image input support</td></tr><tr><td>Gemini</td><td><code>gemini/gemini-live-2.5-flash-preview-native-audio-09-2025</code></td><td>1M</td><td>$0.30</td><td>$2.00</td><td>Native audio, vision, web search</td></tr><tr><td>Vertex AI</td><td><code>vertex_ai/minimaxai/minimax-m2-maas</code></td><td>196K</td><td>$0.30</td><td>$1.20</td><td>Function calling, tool choice</td></tr><tr><td>NVIDIA</td><td><code>nvidia/nemotron-nano-9b-v2</code></td><td>-</td><td>-</td><td>-</td><td>Chat completions</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="ocr-models">OCR Models<a href="https://docs.litellm.ai/release_notes/v1-79-3#ocr-models" class="hash-link" aria-label="Direct link to OCR Models" title="Direct link to OCR Models">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Cost Per Page</th><th>Features</th></tr></thead><tbody><tr><td>Azure AI</td><td><code>azure_ai/doc-intelligence/prebuilt-read</code></td><td>$0.0015</td><td>Document reading</td></tr><tr><td>Azure AI</td><td><code>azure_ai/doc-intelligence/prebuilt-layout</code></td><td>$0.01</td><td>Layout analysis</td></tr><tr><td>Azure AI</td><td><code>azure_ai/doc-intelligence/prebuilt-document</code></td><td>$0.01</td><td>Document processing</td></tr><tr><td>Vertex AI</td><td><code>vertex_ai/mistral-ocr-2505</code></td><td>$0.0005</td><td>OCR processing</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="search-models">Search Models<a href="https://docs.litellm.ai/release_notes/v1-79-3#search-models" class="hash-link" aria-label="Direct link to Search Models" title="Direct link to Search Models">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Pricing</th><th>Features</th></tr></thead><tbody><tr><td>Firecrawl</td><td><code>firecrawl/search</code></td><td>Tiered: $0.00166-$0.0166/query</td><td>10-100 results per query</td></tr><tr><td>SearXNG</td><td><code>searxng/search</code></td><td>Free</td><td>Open-source metasearch</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-79-3#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/azure">Azure</a></strong></p>
<ul>
<li>Add Azure GPT-5-Pro Responses API support with reasoning capabilities - <a href="https://github.com/BerriAI/litellm/pull/16235" target="_blank" rel="noopener noreferrer">PR #16235</a></li>
<li>Add gpt-image-1-mini pricing for Azure with quality tiers (low/medium/high) - <a href="https://github.com/BerriAI/litellm/pull/16182" target="_blank" rel="noopener noreferrer">PR #16182</a></li>
<li>Add support for returning Azure Content Policy error information when exceptions from Azure OpenAI occur - <a href="https://github.com/BerriAI/litellm/pull/16231" target="_blank" rel="noopener noreferrer">PR #16231</a></li>
<li>Fix Azure GPT-5 incorrectly routed to O-series config (temperature parameter unsupported) - <a href="https://github.com/BerriAI/litellm/pull/16246" target="_blank" rel="noopener noreferrer">PR #16246</a></li>
<li>Fix Azure doesn't accept extra body param - <a href="https://github.com/BerriAI/litellm/pull/16116" target="_blank" rel="noopener noreferrer">PR #16116</a></li>
<li>Fix Azure DALL-E-3 health check content policy violation by using safe default prompt - <a href="https://github.com/BerriAI/litellm/pull/16329" target="_blank" rel="noopener noreferrer">PR #16329</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong></p>
<ul>
<li>Fix empty assistant message handling in AWS Bedrock Converse API to prevent 400 Bad Request errors - <a href="https://github.com/BerriAI/litellm/pull/15850" target="_blank" rel="noopener noreferrer">PR #15850</a></li>
<li>Fix: Filter AWS authentication params from Bedrock InvokeModel request body - <a href="https://github.com/BerriAI/litellm/pull/16315" target="_blank" rel="noopener noreferrer">PR #16315</a></li>
<li>Fix Bedrock proxy adding name to file content, breaks when cache_control in use - <a href="https://github.com/BerriAI/litellm/pull/16275" target="_blank" rel="noopener noreferrer">PR #16275</a></li>
<li>Fix global.anthropic.claude-haiku-4-5-20251001-v1:0 supports_reasoning flag and update pricing - <a href="https://github.com/BerriAI/litellm/pull/16263" target="_blank" rel="noopener noreferrer">PR #16263</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Gemini (Google AI Studio + Vertex AI)</a></strong></p>
<ul>
<li>Add gemini live audio model cost in model map - <a href="https://github.com/BerriAI/litellm/pull/16183" target="_blank" rel="noopener noreferrer">PR #16183</a></li>
<li>Fix translation problem with Gemini parallel tool calls - <a href="https://github.com/BerriAI/litellm/pull/16194" target="_blank" rel="noopener noreferrer">PR #16194</a></li>
<li>Fix: Send Gemini API key via x-goog-api-key header with custom api_base - <a href="https://github.com/BerriAI/litellm/pull/16085" target="_blank" rel="noopener noreferrer">PR #16085</a></li>
<li>Fix image_config.aspect_ratio not working for gemini-2.5-flash-image - <a href="https://github.com/BerriAI/litellm/pull/15999" target="_blank" rel="noopener noreferrer">PR #15999</a></li>
<li>Fix Gemini minimal reasoning env overrides disabling thoughts - <a href="https://github.com/BerriAI/litellm/pull/16347" target="_blank" rel="noopener noreferrer">PR #16347</a></li>
<li>Fix cache_read_input_token_cost for gemini-2.5-flash - <a href="https://github.com/BerriAI/litellm/pull/16354" target="_blank" rel="noopener noreferrer">PR #16354</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Fix Anthropic token counting for VertexAI - <a href="https://github.com/BerriAI/litellm/pull/16171" target="_blank" rel="noopener noreferrer">PR #16171</a></li>
<li>Fix anthropic-adapter: properly translate Anthropic image format to OpenAI - <a href="https://github.com/BerriAI/litellm/pull/16202" target="_blank" rel="noopener noreferrer">PR #16202</a></li>
<li>Enable automated prompt caching message format for Claude on Databricks - <a href="https://github.com/BerriAI/litellm/pull/16200" target="_blank" rel="noopener noreferrer">PR #16200</a></li>
<li>Add support for Anthropic Memory Tool - <a href="https://github.com/BerriAI/litellm/pull/16115" target="_blank" rel="noopener noreferrer">PR #16115</a></li>
<li>Propagate cache creation/read token costs for model info to fix Anthropic long context cost calculations - <a href="https://github.com/BerriAI/litellm/pull/16376" target="_blank" rel="noopener noreferrer">PR #16376</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vertex_ai">Vertex AI</a></strong></p>
<ul>
<li>Add Vertex MiniMAX m2 model support - <a href="https://github.com/BerriAI/litellm/pull/16373" target="_blank" rel="noopener noreferrer">PR #16373</a></li>
<li>Correctly map 429 Resource Exhausted to RateLimitError - <a href="https://github.com/BerriAI/litellm/pull/16363" target="_blank" rel="noopener noreferrer">PR #16363</a></li>
<li>Add <code>vertex_credentials</code> support to <code>litellm.rerank()</code> for Vertex AI - <a href="https://github.com/BerriAI/litellm/pull/16266" target="_blank" rel="noopener noreferrer">PR #16266</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/databricks">Databricks</a></strong></p>
<ul>
<li>Fix databricks streaming - <a href="https://github.com/BerriAI/litellm/pull/16368" target="_blank" rel="noopener noreferrer">PR #16368</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/deepgram">Deepgram</a></strong></p>
<ul>
<li>Return the diarized transcript when it's required in the request - <a href="https://github.com/BerriAI/litellm/pull/16133" target="_blank" rel="noopener noreferrer">PR #16133</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/fireworks_ai">Fireworks</a></strong></p>
<ul>
<li>Update Fireworks audio endpoints to new <code>api.fireworks.ai</code> domains - <a href="https://github.com/BerriAI/litellm/pull/16346" target="_blank" rel="noopener noreferrer">PR #16346</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/cohere">Cohere</a></strong></p>
<ul>
<li>Add cohere embed-v4.0 model support - <a href="https://github.com/BerriAI/litellm/pull/16358" target="_blank" rel="noopener noreferrer">PR #16358</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/watsonx">Watsonx</a></strong></p>
<ul>
<li>Support <code>reasoning_effort</code> for watsonx chat models - <a href="https://github.com/BerriAI/litellm/pull/16261" target="_blank" rel="noopener noreferrer">PR #16261</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI</a></strong></p>
<ul>
<li>Remove automatic summary from reasoning_effort transformation - <a href="https://github.com/BerriAI/litellm/pull/16210" target="_blank" rel="noopener noreferrer">PR #16210</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/xai">XAI</a></strong></p>
<ul>
<li>Remove Grok 4 Models Reasoning Effort Parameter - <a href="https://github.com/BerriAI/litellm/pull/16265" target="_blank" rel="noopener noreferrer">PR #16265</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vllm">Hosted VLLM</a></strong></p>
<ul>
<li>Fix HostedVLLMRerankConfig will not be used - <a href="https://github.com/BerriAI/litellm/pull/16352" target="_blank" rel="noopener noreferrer">PR #16352</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-provider-support">New Provider Support<a href="https://docs.litellm.ai/release_notes/v1-79-3#new-provider-support" class="hash-link" aria-label="Direct link to New Provider Support" title="Direct link to New Provider Support">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock Agentcore</a></strong>
<ul>
<li>Add Bedrock Agentcore as a provider on LiteLLM Python SDK and LiteLLM AI Gateway - <a href="https://github.com/BerriAI/litellm/pull/16252" target="_blank" rel="noopener noreferrer">PR #16252</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-79-3#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-79-3#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/ocr">OCR API</a></strong></p>
<ul>
<li>Add VertexAI OCR provider support + cost tracking - <a href="https://github.com/BerriAI/litellm/pull/16216" target="_blank" rel="noopener noreferrer">PR #16216</a></li>
<li>Add Azure AI Doc Intelligence OCR support - <a href="https://github.com/BerriAI/litellm/pull/16219" target="_blank" rel="noopener noreferrer">PR #16219</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/search">Search API</a></strong></p>
<ul>
<li>Add firecrawl search API support with tiered pricing - <a href="https://github.com/BerriAI/litellm/pull/16257" target="_blank" rel="noopener noreferrer">PR #16257</a></li>
<li>Add searxng search API provider - <a href="https://github.com/BerriAI/litellm/pull/16259" target="_blank" rel="noopener noreferrer">PR #16259</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/response_api">Responses API</a></strong></p>
<ul>
<li>Support responses API streaming in langfuse otel - <a href="https://github.com/BerriAI/litellm/pull/16153" target="_blank" rel="noopener noreferrer">PR #16153</a></li>
<li>Pass extra_body parameters to provider in Responses API requests - <a href="https://github.com/BerriAI/litellm/pull/16320" target="_blank" rel="noopener noreferrer">PR #16320</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/container_api">Container API</a></strong></p>
<ul>
<li>Add E2E Container API Support - <a href="https://github.com/BerriAI/litellm/pull/16136" target="_blank" rel="noopener noreferrer">PR #16136</a></li>
<li>Update container documentation to be similar to others - <a href="https://github.com/BerriAI/litellm/pull/16327" target="_blank" rel="noopener noreferrer">PR #16327</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/video_generation">Video Generation API</a></strong></p>
<ul>
<li>Add Vertex and Gemini Videos API with Cost Tracking + UI support - <a href="https://github.com/BerriAI/litellm/pull/16323" target="_blank" rel="noopener noreferrer">PR #16323</a></li>
<li>Add <code>custom_llm_provider</code> support for video endpoints (non-generation) - <a href="https://github.com/BerriAI/litellm/pull/16121" target="_blank" rel="noopener noreferrer">PR #16121</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/audio">Audio API</a></strong></p>
<ul>
<li>Add gpt-4o-transcribe cost tracking - <a href="https://github.com/BerriAI/litellm/pull/16412" target="_blank" rel="noopener noreferrer">PR #16412</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/vector_stores">Vector Stores</a></strong></p>
<ul>
<li>Milvus - search vector store support + support multi-part form data on passthrough - <a href="https://github.com/BerriAI/litellm/pull/16035" target="_blank" rel="noopener noreferrer">PR #16035</a></li>
<li>Azure AI Vector Stores - support "virtual" indexes + create vector store on passthrough API - <a href="https://github.com/BerriAI/litellm/pull/16160" target="_blank" rel="noopener noreferrer">PR #16160</a></li>
<li>Milvus - Passthrough API support - adds create + read vector store support via passthrough API's - <a href="https://github.com/BerriAI/litellm/pull/16170" target="_blank" rel="noopener noreferrer">PR #16170</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/embedding/supported_embedding">Embeddings API</a></strong></p>
<ul>
<li>Use valid CallTypes enum value in embeddings endpoint - <a href="https://github.com/BerriAI/litellm/pull/16328" target="_blank" rel="noopener noreferrer">PR #16328</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/rerank">Rerank API</a></strong></p>
<ul>
<li>Generalize tiered pricing in generic cost calculator - <a href="https://github.com/BerriAI/litellm/pull/16150" target="_blank" rel="noopener noreferrer">PR #16150</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs">Bugs<a href="https://docs.litellm.ai/release_notes/v1-79-3#bugs" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>General</strong>
<ul>
<li>Fix index field not populated in streaming mode with n&gt;1 and tool calls - <a href="https://github.com/BerriAI/litellm/pull/15962" target="_blank" rel="noopener noreferrer">PR #15962</a></li>
<li>Pass aws_region_name in litellm_params - <a href="https://github.com/BerriAI/litellm/pull/16321" target="_blank" rel="noopener noreferrer">PR #16321</a></li>
<li>Add <code>retry-after</code> header support for errors <code>502</code>, <code>503</code>, <code>504</code> - <a href="https://github.com/BerriAI/litellm/pull/16288" target="_blank" rel="noopener noreferrer">PR #16288</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-79-3#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-79-3#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong>Virtual Keys</strong></p>
<ul>
<li>UI - Delete Team Member with friction - <a href="https://github.com/BerriAI/litellm/pull/16167" target="_blank" rel="noopener noreferrer">PR #16167</a></li>
<li>UI - Litellm test key audio support - <a href="https://github.com/BerriAI/litellm/pull/16251" target="_blank" rel="noopener noreferrer">PR #16251</a></li>
<li>UI - Test Key Page Revert Model To Single Select - <a href="https://github.com/BerriAI/litellm/pull/16390" target="_blank" rel="noopener noreferrer">PR #16390</a></li>
</ul>
</li>
<li>
<p><strong>Models + Endpoints</strong></p>
<ul>
<li>UI - Add Model Existing Credentials Improvement - <a href="https://github.com/BerriAI/litellm/pull/16166" target="_blank" rel="noopener noreferrer">PR #16166</a></li>
<li>UI - Add Azure AD Token field and Azure API Key optional - <a href="https://github.com/BerriAI/litellm/pull/16331" target="_blank" rel="noopener noreferrer">PR #16331</a></li>
<li>UI - Fixed Label for vLLM in Model Create Flow - <a href="https://github.com/BerriAI/litellm/pull/16285" target="_blank" rel="noopener noreferrer">PR #16285</a></li>
<li>UI - Include Model Access Group Models on Team Models Table - <a href="https://github.com/BerriAI/litellm/pull/16298" target="_blank" rel="noopener noreferrer">PR #16298</a></li>
<li>Fix /model_group/info Returning Entire Model List for SSO Users - <a href="https://github.com/BerriAI/litellm/pull/16296" target="_blank" rel="noopener noreferrer">PR #16296</a></li>
<li>Litellm non root docker Model Hub Table fix - <a href="https://github.com/BerriAI/litellm/pull/16282" target="_blank" rel="noopener noreferrer">PR #16282</a></li>
</ul>
</li>
<li>
<p><strong>Guardrails</strong></p>
<ul>
<li>UI - Fix regression where Guardrail Entity Could not be selected and entity was not displayed - <a href="https://github.com/BerriAI/litellm/pull/16165" target="_blank" rel="noopener noreferrer">PR #16165</a></li>
<li>UI - Guardrail Info Page Show PII Config - <a href="https://github.com/BerriAI/litellm/pull/16164" target="_blank" rel="noopener noreferrer">PR #16164</a></li>
<li>Change guardrail_information to list type - <a href="https://github.com/BerriAI/litellm/pull/16127" target="_blank" rel="noopener noreferrer">PR #16127</a></li>
<li>UI - LiteLLM Guardrail - ensure you can see UI Friendly name for PII Patterns - <a href="https://github.com/BerriAI/litellm/pull/16382" target="_blank" rel="noopener noreferrer">PR #16382</a></li>
<li>UI - Guardrails - LiteLLM Content Filter, Allow Viewing/Editing Content Filter Settings - <a href="https://github.com/BerriAI/litellm/pull/16383" target="_blank" rel="noopener noreferrer">PR #16383</a></li>
<li>UI - Guardrails - allow updating guardrails through UI. Ensure litellm_params actually get updated in memory - <a href="https://github.com/BerriAI/litellm/pull/16384" target="_blank" rel="noopener noreferrer">PR #16384</a></li>
</ul>
</li>
<li>
<p><strong>SSO Settings</strong></p>
<ul>
<li>Support dot notation on ui sso - <a href="https://github.com/BerriAI/litellm/pull/16135" target="_blank" rel="noopener noreferrer">PR #16135</a></li>
<li>UI - Prevent trailing slash in sso proxy base url input - <a href="https://github.com/BerriAI/litellm/pull/16244" target="_blank" rel="noopener noreferrer">PR #16244</a></li>
<li>UI - SSO Proxy Base URL input validation and remove normalizing / - <a href="https://github.com/BerriAI/litellm/pull/16332" target="_blank" rel="noopener noreferrer">PR #16332</a></li>
<li>UI - Surface SSO Create errors on create flow - <a href="https://github.com/BerriAI/litellm/pull/16369" target="_blank" rel="noopener noreferrer">PR #16369</a></li>
</ul>
</li>
<li>
<p><strong>Usage &amp; Analytics</strong></p>
<ul>
<li>UI - Tag Usage Top Model Table View and Label Fix - <a href="https://github.com/BerriAI/litellm/pull/16249" target="_blank" rel="noopener noreferrer">PR #16249</a></li>
<li>UI - Litellm usage date picker - <a href="https://github.com/BerriAI/litellm/pull/16264" target="_blank" rel="noopener noreferrer">PR #16264</a></li>
</ul>
</li>
<li>
<p><strong>Cache Settings</strong></p>
<ul>
<li>UI - Cache Settings Redis Add Semantic Cache Settings - <a href="https://github.com/BerriAI/litellm/pull/16398" target="_blank" rel="noopener noreferrer">PR #16398</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs-1">Bugs<a href="https://docs.litellm.ai/release_notes/v1-79-3#bugs-1" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>General</strong>
<ul>
<li>UI - Remove encoding_format in request for embedding models - <a href="https://github.com/BerriAI/litellm/pull/16367" target="_blank" rel="noopener noreferrer">PR #16367</a></li>
<li>UI - Revert Changes for Test Key Multiple Model Select - <a href="https://github.com/BerriAI/litellm/pull/16372" target="_blank" rel="noopener noreferrer">PR #16372</a></li>
<li>UI - Various Small Issues - <a href="https://github.com/BerriAI/litellm/pull/16406" target="_blank" rel="noopener noreferrer">PR #16406</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="ai-integrations">AI Integrations<a href="https://docs.litellm.ai/release_notes/v1-79-3#ai-integrations" class="hash-link" aria-label="Direct link to AI Integrations" title="Direct link to AI Integrations">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="logging">Logging<a href="https://docs.litellm.ai/release_notes/v1-79-3#logging" class="hash-link" aria-label="Direct link to Logging" title="Direct link to Logging">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#langfuse">Langfuse</a></strong></p>
<ul>
<li>Fix langfuse input tokens logic for cached tokens - <a href="https://github.com/BerriAI/litellm/pull/16203" target="_blank" rel="noopener noreferrer">PR #16203</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#opik">Opik</a></strong></p>
<ul>
<li>Fix the bug with not incorrect attachment to existing trace &amp; refactor - <a href="https://github.com/BerriAI/litellm/pull/15529" target="_blank" rel="noopener noreferrer">PR #15529</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#s3">S3</a></strong></p>
<ul>
<li>S3 logger, add support for ssl_verify when using minio logger - <a href="https://github.com/BerriAI/litellm/pull/16211" target="_blank" rel="noopener noreferrer">PR #16211</a></li>
<li>Strip base64 in s3 - <a href="https://github.com/BerriAI/litellm/pull/16157" target="_blank" rel="noopener noreferrer">PR #16157</a></li>
<li>Add allowing Key based prefix to s3 path - <a href="https://github.com/BerriAI/litellm/pull/16237" target="_blank" rel="noopener noreferrer">PR #16237</a></li>
<li>Add Prometheus metric to track callback logging failures in S3 - <a href="https://github.com/BerriAI/litellm/pull/16209" target="_blank" rel="noopener noreferrer">PR #16209</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#opentelemetry">OpenTelemetry</a></strong></p>
<ul>
<li>OTEL - Log Cost Breakdown on OTEL Logger - <a href="https://github.com/BerriAI/litellm/pull/16334" target="_blank" rel="noopener noreferrer">PR #16334</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#datadog">DataDog</a></strong></p>
<ul>
<li>Add DD Agent Host support for <code>datadog</code> callback - <a href="https://github.com/BerriAI/litellm/pull/16379" target="_blank" rel="noopener noreferrer">PR #16379</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails">Guardrails<a href="https://docs.litellm.ai/release_notes/v1-79-3#guardrails" class="hash-link" aria-label="Direct link to Guardrails" title="Direct link to Guardrails">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">Noma</a></strong></p>
<ul>
<li>Revert Noma Apply Guardrail implementation - <a href="https://github.com/BerriAI/litellm/pull/16214" target="_blank" rel="noopener noreferrer">PR #16214</a></li>
<li>Litellm noma guardrail support images - <a href="https://github.com/BerriAI/litellm/pull/16199" target="_blank" rel="noopener noreferrer">PR #16199</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">PANW Prisma AIRS</a></strong></p>
<ul>
<li>PANW prisma airs guardrail deduplication and enhanced session tracking - <a href="https://github.com/BerriAI/litellm/pull/16273" target="_blank" rel="noopener noreferrer">PR #16273</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">LiteLLM Custom Guardrail</a></strong></p>
<ul>
<li>Add LiteLLM Gateway built in guardrail - <a href="https://github.com/BerriAI/litellm/pull/16338" target="_blank" rel="noopener noreferrer">PR #16338</a></li>
<li>UI - Allow configuring LiteLLM Custom Guardrail - <a href="https://github.com/BerriAI/litellm/pull/16339" target="_blank" rel="noopener noreferrer">PR #16339</a></li>
<li>Bug Fix: Content Filter Guard - <a href="https://github.com/BerriAI/litellm/pull/16414" target="_blank" rel="noopener noreferrer">PR #16414</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="secret-managers">Secret Managers<a href="https://docs.litellm.ai/release_notes/v1-79-3#secret-managers" class="hash-link" aria-label="Direct link to Secret Managers" title="Direct link to Secret Managers">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/secret_managers">CyberArk</a></strong></p>
<ul>
<li>Add CyberArk Secrets Manager Integration - <a href="https://github.com/BerriAI/litellm/pull/16278" target="_blank" rel="noopener noreferrer">PR #16278</a></li>
<li>Cyber Ark - Add Key Rotations support - <a href="https://github.com/BerriAI/litellm/pull/16289" target="_blank" rel="noopener noreferrer">PR #16289</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/secret_managers">HashiCorp Vault</a></strong></p>
<ul>
<li>Add configurable mount name and path prefix for HashiCorp Vault - <a href="https://github.com/BerriAI/litellm/pull/16253" target="_blank" rel="noopener noreferrer">PR #16253</a></li>
<li>Secret Manager - Hashicorp, add auth via approle - <a href="https://github.com/BerriAI/litellm/pull/16374" target="_blank" rel="noopener noreferrer">PR #16374</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/secret_managers">AWS Secrets Manager</a></strong></p>
<ul>
<li>Add tags and descriptions support to aws secrets manager - <a href="https://github.com/BerriAI/litellm/pull/16224" target="_blank" rel="noopener noreferrer">PR #16224</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/secret_managers">Custom Secret Manager</a></strong></p>
<ul>
<li>Add Custom Secret Manager - Allow users to define and write a custom secret manager - <a href="https://github.com/BerriAI/litellm/pull/16297" target="_blank" rel="noopener noreferrer">PR #16297</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Email Notifications - Ensure Users get Key Rotated Email - <a href="https://github.com/BerriAI/litellm/pull/16292" target="_blank" rel="noopener noreferrer">PR #16292</a></li>
<li>Fix verify ssl on sts boto3 - <a href="https://github.com/BerriAI/litellm/pull/16313" target="_blank" rel="noopener noreferrer">PR #16313</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-79-3#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li><strong>Cost Tracking</strong>
<ul>
<li>Fix OpenAI Responses API streaming tests usage field names and cost calculation - <a href="https://github.com/BerriAI/litellm/pull/16236" target="_blank" rel="noopener noreferrer">PR #16236</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-gateway">MCP Gateway<a href="https://docs.litellm.ai/release_notes/v1-79-3#mcp-gateway" class="hash-link" aria-label="Direct link to MCP Gateway" title="Direct link to MCP Gateway">​</a></h2>
<ul>
<li><strong>Configuration</strong>
<ul>
<li>Configure static mcp header - <a href="https://github.com/BerriAI/litellm/pull/16179" target="_blank" rel="noopener noreferrer">PR #16179</a></li>
<li>Persist mcp credentials in db - <a href="https://github.com/BerriAI/litellm/pull/16308" target="_blank" rel="noopener noreferrer">PR #16308</a></li>
</ul>
</li>
</ul>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements">Performance / Loadbalancing / Reliability improvements<a href="https://docs.litellm.ai/release_notes/v1-79-3#performance--loadbalancing--reliability-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements" title="Direct link to Performance / Loadbalancing / Reliability improvements">​</a></h2>
<ul>
<li>
<p><strong>Memory Leak Fixes</strong></p>
<ul>
<li>Resolve memory accumulation caused by Pydantic 2.11+ deprecation warnings - <a href="https://github.com/BerriAI/litellm/pull/16110" target="_blank" rel="noopener noreferrer">PR #16110</a></li>
</ul>
</li>
<li>
<p><strong>Session Management</strong></p>
<ul>
<li>Add shared_session support to responses API - <a href="https://github.com/BerriAI/litellm/pull/16260" target="_blank" rel="noopener noreferrer">PR #16260</a></li>
</ul>
</li>
<li>
<p><strong>Error Handling</strong></p>
<ul>
<li>Gracefully handle connection closed errors during streaming - <a href="https://github.com/BerriAI/litellm/pull/16294" target="_blank" rel="noopener noreferrer">PR #16294</a></li>
<li>Handle None values in daily spend sort key - <a href="https://github.com/BerriAI/litellm/pull/16245" target="_blank" rel="noopener noreferrer">PR #16245</a></li>
</ul>
</li>
<li>
<p><strong>Configuration</strong></p>
<ul>
<li>Remove minimum validation for cache control injection index - <a href="https://github.com/BerriAI/litellm/pull/16149" target="_blank" rel="noopener noreferrer">PR #16149</a></li>
<li>Improve clearing logic - only remove unvisited endpoints - <a href="https://github.com/BerriAI/litellm/pull/16400" target="_blank" rel="noopener noreferrer">PR #16400</a></li>
</ul>
</li>
<li>
<p><strong>Redis</strong></p>
<ul>
<li>Handle float redis_version from AWS ElastiCache Valkey - <a href="https://github.com/BerriAI/litellm/pull/16207" target="_blank" rel="noopener noreferrer">PR #16207</a></li>
</ul>
</li>
<li>
<p><strong>Hooks</strong></p>
<ul>
<li>Add parallel execution handling in during_call_hook - <a href="https://github.com/BerriAI/litellm/pull/16279" target="_blank" rel="noopener noreferrer">PR #16279</a></li>
</ul>
</li>
<li>
<p><strong>Infrastructure</strong></p>
<ul>
<li>Install runtime node for prisma - <a href="https://github.com/BerriAI/litellm/pull/16410" target="_blank" rel="noopener noreferrer">PR #16410</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates">Documentation Updates<a href="https://docs.litellm.ai/release_notes/v1-79-3#documentation-updates" class="hash-link" aria-label="Direct link to Documentation Updates" title="Direct link to Documentation Updates">​</a></h2>
<ul>
<li>
<p><strong>Provider Documentation</strong></p>
<ul>
<li>Docs - v1.79.1 - <a href="https://github.com/BerriAI/litellm/pull/16163" target="_blank" rel="noopener noreferrer">PR #16163</a></li>
<li>Fix broken link on model_management.md - <a href="https://github.com/BerriAI/litellm/pull/16217" target="_blank" rel="noopener noreferrer">PR #16217</a></li>
<li>Fix image generation response format - use 'images' array instead of 'image' object - <a href="https://github.com/BerriAI/litellm/pull/16378" target="_blank" rel="noopener noreferrer">PR #16378</a></li>
</ul>
</li>
<li>
<p><strong>General Documentation</strong></p>
<ul>
<li>Add minimum resource requirement for production - <a href="https://github.com/BerriAI/litellm/pull/16146" target="_blank" rel="noopener noreferrer">PR #16146</a></li>
<li>Add benchmark comparison with other AI gateways - <a href="https://github.com/BerriAI/litellm/pull/16248" target="_blank" rel="noopener noreferrer">PR #16248</a></li>
<li>LiteLLM content filter guard documentation - <a href="https://github.com/BerriAI/litellm/pull/16413" target="_blank" rel="noopener noreferrer">PR #16413</a></li>
<li>Fix typo of the word orginal - <a href="https://github.com/BerriAI/litellm/pull/16255" target="_blank" rel="noopener noreferrer">PR #16255</a></li>
</ul>
</li>
<li>
<p><strong>Security</strong></p>
<ul>
<li>Remove tornado test files (including test.key), fixes Python 3.13 security issues - <a href="https://github.com/BerriAI/litellm/pull/16342" target="_blank" rel="noopener noreferrer">PR #16342</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-79-3#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@steve-gore-snapdocs made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16149" target="_blank" rel="noopener noreferrer">PR #16149</a></li>
<li>@timbmg made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16120" target="_blank" rel="noopener noreferrer">PR #16120</a></li>
<li>@Nivg made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16202" target="_blank" rel="noopener noreferrer">PR #16202</a></li>
<li>@pablobgar made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16194" target="_blank" rel="noopener noreferrer">PR #16194</a></li>
<li>@AlanPonnachan made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16150" target="_blank" rel="noopener noreferrer">PR #16150</a></li>
<li>@Chesars made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16236" target="_blank" rel="noopener noreferrer">PR #16236</a></li>
<li>@bowenliang123 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16255" target="_blank" rel="noopener noreferrer">PR #16255</a></li>
<li>@dean-zavad made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16199" target="_blank" rel="noopener noreferrer">PR #16199</a></li>
<li>@alexkuzmik made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15529" target="_blank" rel="noopener noreferrer">PR #15529</a></li>
<li>@Granine made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16281" target="_blank" rel="noopener noreferrer">PR #16281</a></li>
<li>@Oodapow made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16279" target="_blank" rel="noopener noreferrer">PR #16279</a></li>
<li>@jgoodyear made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16275" target="_blank" rel="noopener noreferrer">PR #16275</a></li>
<li>@Qanpi made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16321" target="_blank" rel="noopener noreferrer">PR #16321</a></li>
<li>@ShimonMimoun made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16313" target="_blank" rel="noopener noreferrer">PR #16313</a></li>
<li>@andriykislitsyn made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16288" target="_blank" rel="noopener noreferrer">PR #16288</a></li>
<li>@reckless-huang made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16263" target="_blank" rel="noopener noreferrer">PR #16263</a></li>
<li>@chenmoneygithub made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16368" target="_blank" rel="noopener noreferrer">PR #16368</a></li>
<li>@stembe-digitalex made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16354" target="_blank" rel="noopener noreferrer">PR #16354</a></li>
<li>@jfcherng made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16352" target="_blank" rel="noopener noreferrer">PR #16352</a></li>
<li>@xingyaoww made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16246" target="_blank" rel="noopener noreferrer">PR #16246</a></li>
<li>@emerzon made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16373" target="_blank" rel="noopener noreferrer">PR #16373</a></li>
<li>@wwwillchen made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16376" target="_blank" rel="noopener noreferrer">PR #16376</a></li>
<li>@fabriciojoc made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16203" target="_blank" rel="noopener noreferrer">PR #16203</a></li>
<li>@jroberts2600 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16273" target="_blank" rel="noopener noreferrer">PR #16273</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog">Full Changelog<a href="https://docs.litellm.ai/release_notes/v1-79-3#full-changelog" class="hash-link" aria-label="Direct link to Full Changelog" title="Direct link to Full Changelog">​</a></h2>
<p><strong><a href="https://github.com/BerriAI/litellm/compare/v1.79.1-nightly...v1.79.2.rc.1" target="_blank" rel="noopener noreferrer">View complete changelog on GitHub</a></strong></p>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[v1.79.1-stable - Guardrail Playground]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-79-1</id>
        <link href="https://docs.litellm.ai/release_notes/v1-79-1"/>
        <updated>2025-11-01T10:00:00.000Z</updated>
        <summary type="html"><![CDATA[Deploy this version]]></summary>
        <content type="html"><![CDATA[<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-79-1#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">docker.litellm.ai/berriai/litellm:v1.79.1-stable</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.80.0</span><br></span></code></pre></div></div></div></div></div>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-79-1#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<ul>
<li><strong>Container API Support</strong> - End-to-end OpenAI Container API support with proxy integration, logging, and cost tracking</li>
<li><strong>FAL AI Image Generation</strong> - Native support for FAL AI image generation models with cost tracking</li>
<li><strong>UI Enhancements</strong> - Guardrail Playground, Cache Settings, Tag Routing, SSO Settings</li>
<li><strong>Batch API Rate Limiting</strong> - Input-based rate limits support for Batch API requests</li>
<li><strong>Vector Store Expansion</strong> - Milvus vector store support and Azure AI virtual indexes</li>
<li><strong>Memory Leak Fixes</strong> - Resolved issues accounting for 90% of memory leaks on Python SDK &amp; AI Gateway</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="dependency-upgrades">Dependency Upgrades<a href="https://docs.litellm.ai/release_notes/v1-79-1#dependency-upgrades" class="hash-link" aria-label="Direct link to Dependency Upgrades" title="Direct link to Dependency Upgrades">​</a></h2>
<ul>
<li><strong>Dependencies</strong>
<ul>
<li>Build(deps): bump starlette from 0.47.2 to 0.49.1 - <a href="https://github.com/BerriAI/litellm/pull/16027" target="_blank" rel="noopener noreferrer">PR #16027</a></li>
<li>Build(deps): bump fastapi from 0.116.1 to 0.120.1 - <a href="https://github.com/BerriAI/litellm/pull/16054" target="_blank" rel="noopener noreferrer">PR #16054</a></li>
<li>Build(deps): bump hono from 4.9.7 to 4.10.3 in /litellm-js/spend-logs - <a href="https://github.com/BerriAI/litellm/pull/15915" target="_blank" rel="noopener noreferrer">PR #15915</a></li>
</ul>
</li>
</ul>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-79-1#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support">New Model Support<a href="https://docs.litellm.ai/release_notes/v1-79-1#new-model-support" class="hash-link" aria-label="Direct link to New Model Support" title="Direct link to New Model Support">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th><th>Features</th></tr></thead><tbody><tr><td>Mistral</td><td><code>mistral/codestral-embed</code></td><td>8K</td><td>$0.15</td><td>-</td><td>Embeddings</td></tr><tr><td>Mistral</td><td><code>mistral/codestral-embed-2505</code></td><td>8K</td><td>$0.15</td><td>-</td><td>Embeddings</td></tr><tr><td>Gemini</td><td><code>gemini/gemini-embedding-001</code></td><td>2K</td><td>$0.15</td><td>-</td><td>Embeddings</td></tr><tr><td>FAL AI</td><td><code>fal_ai/fal-ai/flux-pro/v1.1-ultra</code></td><td>-</td><td>-</td><td>-</td><td>Image generation - $0.0398/image</td></tr><tr><td>FAL AI</td><td><code>fal_ai/fal-ai/imagen4/preview</code></td><td>-</td><td>-</td><td>-</td><td>Image generation - $0.0398/image</td></tr><tr><td>FAL AI</td><td><code>fal_ai/fal-ai/recraft/v3/text-to-image</code></td><td>-</td><td>-</td><td>-</td><td>Image generation - $0.0398/image</td></tr><tr><td>FAL AI</td><td><code>fal_ai/fal-ai/stable-diffusion-v35-medium</code></td><td>-</td><td>-</td><td>-</td><td>Image generation - $0.0398/image</td></tr><tr><td>FAL AI</td><td><code>fal_ai/bria/text-to-image/3.2</code></td><td>-</td><td>-</td><td>-</td><td>Image generation - $0.0398/image</td></tr><tr><td>OpenAI</td><td><code>openai/sora-2-pro</code></td><td>-</td><td>-</td><td>-</td><td>Video generation - $0.30/video/second</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-79-1#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Extended Claude 3-7 Sonnet deprecation date from 2026-02-01 to 2026-02-19 - <a href="https://github.com/BerriAI/litellm/pull/15976" target="_blank" rel="noopener noreferrer">PR #15976</a></li>
<li>Extended Claude Opus 4-0 deprecation date from 2025-03-01 to 2026-05-01 - <a href="https://github.com/BerriAI/litellm/pull/15976" target="_blank" rel="noopener noreferrer">PR #15976</a></li>
<li>Removed Claude Haiku 3-5 deprecation date (previously 2025-03-01) - <a href="https://github.com/BerriAI/litellm/pull/15976" target="_blank" rel="noopener noreferrer">PR #15976</a></li>
<li>Added Claude Opus 4-1, Claude Opus 4-0 20250513, Claude Sonnet 4 20250514 deprecation dates - <a href="https://github.com/BerriAI/litellm/pull/15976" target="_blank" rel="noopener noreferrer">PR #15976</a></li>
<li>Added web search support for Claude Opus 4-1 - <a href="https://github.com/BerriAI/litellm/pull/15976" target="_blank" rel="noopener noreferrer">PR #15976</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong></p>
<ul>
<li>Fix empty assistant message handling in AWS Bedrock Converse API to prevent 400 Bad Request errors - <a href="https://github.com/BerriAI/litellm/pull/15850" target="_blank" rel="noopener noreferrer">PR #15850</a></li>
<li>Allow using ARNs when generating images via Bedrock - <a href="https://github.com/BerriAI/litellm/pull/15789" target="_blank" rel="noopener noreferrer">PR #15789</a></li>
<li>Add per model group header forwarding for Bedrock Invoke API - <a href="https://github.com/BerriAI/litellm/pull/16042" target="_blank" rel="noopener noreferrer">PR #16042</a></li>
<li>Preserve Bedrock inference profile IDs in health checks - <a href="https://github.com/BerriAI/litellm/pull/15947" target="_blank" rel="noopener noreferrer">PR #15947</a></li>
<li>Added fallback logic for detecting file content-type when S3 returns generic type - When using Bedrock with S3-hosted files, if the S3 object's Content-Type is not correctly set (e.g., binary/octet-stream instead of image/png), Bedrock can now handle it correctly - <a href="https://github.com/BerriAI/litellm/pull/15635" target="_blank" rel="noopener noreferrer">PR #15635</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/azure">Azure</a></strong></p>
<ul>
<li>Add deprecation dates for Azure OpenAI models (gpt-4o-2024-08-06, gpt-4o-2024-11-20, gpt-4.1 series, o3-2025-04-16, text-embedding-3-small) - <a href="https://github.com/BerriAI/litellm/pull/15976" target="_blank" rel="noopener noreferrer">PR #15976</a></li>
<li>Fix Azure OpenAI ContextWindowExceededError mapping from Azure errors - <a href="https://github.com/BerriAI/litellm/pull/15981" target="_blank" rel="noopener noreferrer">PR #15981</a></li>
<li>Add handling for <code>v1</code> under Azure API versions - <a href="https://github.com/BerriAI/litellm/pull/15984" target="_blank" rel="noopener noreferrer">PR #15984</a></li>
<li>Fix azure doesn't accept extra body param - <a href="https://github.com/BerriAI/litellm/pull/16116" target="_blank" rel="noopener noreferrer">PR #16116</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI</a></strong></p>
<ul>
<li>Add deprecation dates for gpt-3.5-turbo-1106, gpt-4-0125-preview, gpt-4-1106-preview, o1-mini-2024-09-12 - <a href="https://github.com/BerriAI/litellm/pull/15976" target="_blank" rel="noopener noreferrer">PR #15976</a></li>
<li>Add extended Sora-2 modality support (text + image inputs) - <a href="https://github.com/BerriAI/litellm/pull/15976" target="_blank" rel="noopener noreferrer">PR #15976</a></li>
<li>Updated OpenAI Sora-2-Pro pricing to $0.30/video/second - <a href="https://github.com/BerriAI/litellm/pull/15976" target="_blank" rel="noopener noreferrer">PR #15976</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openrouter">OpenRouter</a></strong></p>
<ul>
<li>Add Claude Haiku 4.5 pricing for OpenRouter - <a href="https://github.com/BerriAI/litellm/pull/15909" target="_blank" rel="noopener noreferrer">PR #15909</a></li>
<li>Add base_url config with environment variables documentation - <a href="https://github.com/BerriAI/litellm/pull/15946" target="_blank" rel="noopener noreferrer">PR #15946</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/mistral">Mistral</a></strong></p>
<ul>
<li>Add codestral-embed-2505 embedding model - <a href="https://github.com/BerriAI/litellm/pull/16071" target="_blank" rel="noopener noreferrer">PR #16071</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Gemini (Google AI Studio + Vertex AI)</a></strong></p>
<ul>
<li>Fix gemini request mutation for tool use - <a href="https://github.com/BerriAI/litellm/pull/16002" target="_blank" rel="noopener noreferrer">PR #16002</a></li>
<li>Add gemini-embedding-001 pricing entry for Google GenAI API - <a href="https://github.com/BerriAI/litellm/pull/16078" target="_blank" rel="noopener noreferrer">PR #16078</a></li>
<li>Changes to fix frequency_penalty and presence_penalty issue for gemini-2.5-pro model - <a href="https://github.com/BerriAI/litellm/pull/16041" target="_blank" rel="noopener noreferrer">PR #16041</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/deepinfra">DeepInfra</a></strong></p>
<ul>
<li>Add vision support for Qwen/Qwen3-chat-32b model - <a href="https://github.com/BerriAI/litellm/pull/15976" target="_blank" rel="noopener noreferrer">PR #15976</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vercel_ai_gateway">Vercel AI Gateway</a></strong></p>
<ul>
<li>Fix vercel_ai_gateway entry for glm-4.6 (moved from vercel_ai_gateway/glm-4.6 to vercel_ai_gateway/zai/glm-4.6) - <a href="https://github.com/BerriAI/litellm/pull/16084" target="_blank" rel="noopener noreferrer">PR #16084</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/fireworks_ai">Fireworks</a></strong></p>
<ul>
<li>Don't add "accounts/fireworks/models" prefix for Fireworks Provider - <a href="https://github.com/BerriAI/litellm/pull/15938" target="_blank" rel="noopener noreferrer">PR #15938</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/cohere">Cohere</a></strong></p>
<ul>
<li>Add OpenAI-compatible annotations support for Cohere v2 citations - <a href="https://github.com/BerriAI/litellm/pull/16038" target="_blank" rel="noopener noreferrer">PR #16038</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/deepgram">Deepgram</a></strong></p>
<ul>
<li>Handle Deepgram detected language when available - <a href="https://github.com/BerriAI/litellm/pull/16093" target="_blank" rel="noopener noreferrer">PR #16093</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-79-1#bug-fixes" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/providers/xai">Xai</a></strong>
<ul>
<li>Add Xai websearch cost tracking - <a href="https://github.com/BerriAI/litellm/pull/16001" target="_blank" rel="noopener noreferrer">PR #16001</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-provider-support">New Provider Support<a href="https://docs.litellm.ai/release_notes/v1-79-1#new-provider-support" class="hash-link" aria-label="Direct link to New Provider Support" title="Direct link to New Provider Support">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/image_generation">FAL AI</a></strong></p>
<ul>
<li>Add FAL AI Image Generation support - <a href="https://github.com/BerriAI/litellm/pull/16067" target="_blank" rel="noopener noreferrer">PR #16067</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/oci">OCI (Oracle Cloud Infrastructure)</a></strong></p>
<ul>
<li>Add OCI Signer Authentication support - <a href="https://github.com/BerriAI/litellm/pull/16064" target="_blank" rel="noopener noreferrer">PR #16064</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-79-1#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-79-1#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/containers">Container API</a></strong></p>
<ul>
<li>Add end-to-end OpenAI Container API support to LiteLLM SDK - <a href="https://github.com/BerriAI/litellm/pull/16136" target="_blank" rel="noopener noreferrer">PR #16136</a></li>
<li>Add proxy support for container APIs - <a href="https://github.com/BerriAI/litellm/pull/16049" target="_blank" rel="noopener noreferrer">PR #16049</a></li>
<li>Add logging support for Container API - <a href="https://github.com/BerriAI/litellm/pull/16049" target="_blank" rel="noopener noreferrer">PR #16049</a></li>
<li>Add cost tracking support for containers with documentation - <a href="https://github.com/BerriAI/litellm/pull/16117" target="_blank" rel="noopener noreferrer">PR #16117</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/response_api">Responses API</a></strong></p>
<ul>
<li>Respect <code>LiteLLM-Disable-Message-Redaction</code> header for Responses API - <a href="https://github.com/BerriAI/litellm/pull/15966" target="_blank" rel="noopener noreferrer">PR #15966</a></li>
<li>Add /openai routes for responses API (Azure OpenAI SDK Compatibility) - <a href="https://github.com/BerriAI/litellm/pull/15988" target="_blank" rel="noopener noreferrer">PR #15988</a></li>
<li>Redact reasoning summaries in ResponsesAPI output when message logging is disabled - <a href="https://github.com/BerriAI/litellm/pull/15965" target="_blank" rel="noopener noreferrer">PR #15965</a></li>
<li>Support text.format parameter in Responses API for providers without native ResponsesAPIConfig - <a href="https://github.com/BerriAI/litellm/pull/16023" target="_blank" rel="noopener noreferrer">PR #16023</a></li>
<li>Add LLM provider response headers to Responses API - <a href="https://github.com/BerriAI/litellm/pull/16091" target="_blank" rel="noopener noreferrer">PR #16091</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/video_generation">Video Generation API</a></strong></p>
<ul>
<li>Add <code>custom_llm_provider</code> support for video endpoints (non-generation) - <a href="https://github.com/BerriAI/litellm/pull/16121" target="_blank" rel="noopener noreferrer">PR #16121</a></li>
<li>Fix documentation for videos - <a href="https://github.com/BerriAI/litellm/pull/15937" target="_blank" rel="noopener noreferrer">PR #15937</a></li>
<li>Add OpenAI client usage documentation for videos and fix navigation visibility - <a href="https://github.com/BerriAI/litellm/pull/15996" target="_blank" rel="noopener noreferrer">PR #15996</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/moderations">Moderations API</a></strong></p>
<ul>
<li>Moderations endpoint now respects <code>api_base</code> configuration parameter - <a href="https://github.com/BerriAI/litellm/pull/16087" target="_blank" rel="noopener noreferrer">PR #16087</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/vector_stores">Vector Stores</a></strong></p>
<ul>
<li>Milvus - search vector store support - <a href="https://github.com/BerriAI/litellm/pull/16035" target="_blank" rel="noopener noreferrer">PR #16035</a></li>
<li>Azure AI Vector Stores - support "virtual" indexes + create vector store on passthrough API - <a href="https://github.com/BerriAI/litellm/pull/16160" target="_blank" rel="noopener noreferrer">PR #16160</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/pass_through/vertex_ai">Passthrough Endpoints</a></strong></p>
<ul>
<li>Support multi-part form data on passthrough - <a href="https://github.com/BerriAI/litellm/pull/16035" target="_blank" rel="noopener noreferrer">PR #16035</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-79-1#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-79-1#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong>Virtual Keys</strong></p>
<ul>
<li>Validation for Proxy Base URL in SSO Settings - <a href="https://github.com/BerriAI/litellm/pull/16082" target="_blank" rel="noopener noreferrer">PR #16082</a></li>
<li>Test Key UI Embeddings support - <a href="https://github.com/BerriAI/litellm/pull/16065" target="_blank" rel="noopener noreferrer">PR #16065</a></li>
<li>Add Key Type Select in Key Settings - <a href="https://github.com/BerriAI/litellm/pull/16034" target="_blank" rel="noopener noreferrer">PR #16034</a></li>
<li>Key Already Exist Error Notification - <a href="https://github.com/BerriAI/litellm/pull/15993" target="_blank" rel="noopener noreferrer">PR #15993</a></li>
</ul>
</li>
<li>
<p><strong>Models + Endpoints</strong></p>
<ul>
<li>Changed API Base from Select to Input in New LLM Credentials - <a href="https://github.com/BerriAI/litellm/pull/15987" target="_blank" rel="noopener noreferrer">PR #15987</a></li>
<li>Remove limit from admin UI numerical input - <a href="https://github.com/BerriAI/litellm/pull/15991" target="_blank" rel="noopener noreferrer">PR #15991</a></li>
<li>Config Models should not be editable - <a href="https://github.com/BerriAI/litellm/pull/16020" target="_blank" rel="noopener noreferrer">PR #16020</a></li>
<li>Add tags in model creation - <a href="https://github.com/BerriAI/litellm/pull/16138" target="_blank" rel="noopener noreferrer">PR #16138</a></li>
<li>Add Tags to update model - <a href="https://github.com/BerriAI/litellm/pull/16140" target="_blank" rel="noopener noreferrer">PR #16140</a></li>
</ul>
</li>
<li>
<p><strong>Guardrails</strong></p>
<ul>
<li>Add Apply Guardrail Testing Playground - <a href="https://github.com/BerriAI/litellm/pull/16030" target="_blank" rel="noopener noreferrer">PR #16030</a></li>
<li>Config Guardrails should not be editable and guardrail info fix - <a href="https://github.com/BerriAI/litellm/pull/16142" target="_blank" rel="noopener noreferrer">PR #16142</a></li>
</ul>
</li>
<li>
<p><strong>Cache Settings</strong></p>
<ul>
<li>Allow setting cache settings on UI - <a href="https://github.com/BerriAI/litellm/pull/16143" target="_blank" rel="noopener noreferrer">PR #16143</a></li>
</ul>
</li>
<li>
<p><strong>Routing</strong></p>
<ul>
<li>Allow setting all routing strategies, tag filtering on UI - <a href="https://github.com/BerriAI/litellm/pull/16139" target="_blank" rel="noopener noreferrer">PR #16139</a></li>
</ul>
</li>
<li>
<p><strong>Admin Settings</strong></p>
<ul>
<li>Add license metadata to health/readiness endpoint - <a href="https://github.com/BerriAI/litellm/pull/15997" target="_blank" rel="noopener noreferrer">PR #15997</a></li>
<li>Litellm Backend SSO Changes - <a href="https://github.com/BerriAI/litellm/pull/16029" target="_blank" rel="noopener noreferrer">PR #16029</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="logging--guardrail--prompt-management-integrations">Logging / Guardrail / Prompt Management Integrations<a href="https://docs.litellm.ai/release_notes/v1-79-1#logging--guardrail--prompt-management-integrations" class="hash-link" aria-label="Direct link to Logging / Guardrail / Prompt Management Integrations" title="Direct link to Logging / Guardrail / Prompt Management Integrations">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-3">Features<a href="https://docs.litellm.ai/release_notes/v1-79-1#features-3" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#opentelemetry">OpenTelemetry</a></strong></p>
<ul>
<li>Enable OpenTelemetry context propagation by external tracers - <a href="https://github.com/BerriAI/litellm/pull/15940" target="_blank" rel="noopener noreferrer">PR #15940</a></li>
<li>Ensure error information is logged on OTEL - <a href="https://github.com/BerriAI/litellm/pull/15978" target="_blank" rel="noopener noreferrer">PR #15978</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#langfuse">Langfuse</a></strong></p>
<ul>
<li>Fix duplicate trace in langfuse_otel - <a href="https://github.com/BerriAI/litellm/pull/15931" target="_blank" rel="noopener noreferrer">PR #15931</a></li>
<li>Support tool usage messages with Langfuse OTEL integration - <a href="https://github.com/BerriAI/litellm/pull/15932" target="_blank" rel="noopener noreferrer">PR #15932</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#datadog">DataDog</a></strong></p>
<ul>
<li>Ensure key's metadata + guardrail is logged on DD - <a href="https://github.com/BerriAI/litellm/pull/15980" target="_blank" rel="noopener noreferrer">PR #15980</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#opik">Opik</a></strong></p>
<ul>
<li>Enhance requester metadata retrieval from API key auth - <a href="https://github.com/BerriAI/litellm/pull/15897" target="_blank" rel="noopener noreferrer">PR #15897</a></li>
<li>User auth key metadata Documentation - <a href="https://github.com/BerriAI/litellm/pull/16004" target="_blank" rel="noopener noreferrer">PR #16004</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#sqs">SQS</a></strong></p>
<ul>
<li>Add Base64 handling for SQS Logger - <a href="https://github.com/BerriAI/litellm/pull/16028" target="_blank" rel="noopener noreferrer">PR #16028</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Fix: User API key and team id and user id missing from custom callback is not misfiring - <a href="https://github.com/BerriAI/litellm/pull/15982" target="_blank" rel="noopener noreferrer">PR #15982</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails">Guardrails<a href="https://docs.litellm.ai/release_notes/v1-79-1#guardrails" class="hash-link" aria-label="Direct link to Guardrails" title="Direct link to Guardrails">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">IBM Guardrails</a></strong></p>
<ul>
<li>Update IBM Guardrails to correctly use SSL Verify argument - <a href="https://github.com/BerriAI/litellm/pull/15975" target="_blank" rel="noopener noreferrer">PR #15975</a></li>
<li>Add additional detail to ibm_guardrails.md documentation - <a href="https://github.com/BerriAI/litellm/pull/15971" target="_blank" rel="noopener noreferrer">PR #15971</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">Model Armor</a></strong></p>
<ul>
<li>Support during_call for model armor guardrails - <a href="https://github.com/BerriAI/litellm/pull/15970" target="_blank" rel="noopener noreferrer">PR #15970</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">Lasso Security</a></strong></p>
<ul>
<li>Upgrade to Lasso API v3 and fix ULID generation - <a href="https://github.com/BerriAI/litellm/pull/15941" target="_blank" rel="noopener noreferrer">PR #15941</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">PANW Prisma AIRS</a></strong></p>
<ul>
<li>Add per-request profile overrides to PANW Prisma AIRS - <a href="https://github.com/BerriAI/litellm/pull/16069" target="_blank" rel="noopener noreferrer">PR #16069</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">Grayswan</a></strong></p>
<ul>
<li>Improve Grayswan guardrail documentation - <a href="https://github.com/BerriAI/litellm/pull/15875" target="_blank" rel="noopener noreferrer">PR #15875</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">Pillar AI</a></strong></p>
<ul>
<li>Graceful degradation for pillar service when using litellm - <a href="https://github.com/BerriAI/litellm/pull/15857" target="_blank" rel="noopener noreferrer">PR #15857</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Ensure Key Guardrails are applied - <a href="https://github.com/BerriAI/litellm/pull/16025" target="_blank" rel="noopener noreferrer">PR #16025</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="prompt-management">Prompt Management<a href="https://docs.litellm.ai/release_notes/v1-79-1#prompt-management" class="hash-link" aria-label="Direct link to Prompt Management" title="Direct link to Prompt Management">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/prompt_management">GitLab</a></strong>
<ul>
<li>Add GitlabPromptCache and enable subfolder access - <a href="https://github.com/BerriAI/litellm/pull/15712" target="_blank" rel="noopener noreferrer">PR #15712</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-79-1#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li>
<p><strong>Cost Tracking</strong></p>
<ul>
<li>Fix spend tracking for OCR/aOCR requests (log <code>pages_processed</code> + recognize <code>OCRResponse</code>) - <a href="https://github.com/BerriAI/litellm/pull/16070" target="_blank" rel="noopener noreferrer">PR #16070</a></li>
</ul>
</li>
<li>
<p><strong>Rate Limiting</strong></p>
<ul>
<li>Add support for Batch API Rate limiting - PR1 adds support for input based rate limits - <a href="https://github.com/BerriAI/litellm/pull/16075" target="_blank" rel="noopener noreferrer">PR #16075</a></li>
<li>Handle multiple rate limit types per descriptor and prevent IndexError - <a href="https://github.com/BerriAI/litellm/pull/16039" target="_blank" rel="noopener noreferrer">PR #16039</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-gateway">MCP Gateway<a href="https://docs.litellm.ai/release_notes/v1-79-1#mcp-gateway" class="hash-link" aria-label="Direct link to MCP Gateway" title="Direct link to MCP Gateway">​</a></h2>
<ul>
<li><strong>OAuth</strong>
<ul>
<li>Add support for dynamic client registration - <a href="https://github.com/BerriAI/litellm/pull/15921" target="_blank" rel="noopener noreferrer">PR #15921</a></li>
<li>Respect X-Forwarded- headers in OAuth endpoints - <a href="https://github.com/BerriAI/litellm/pull/16036" target="_blank" rel="noopener noreferrer">PR #16036</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements">Performance / Loadbalancing / Reliability improvements<a href="https://docs.litellm.ai/release_notes/v1-79-1#performance--loadbalancing--reliability-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements" title="Direct link to Performance / Loadbalancing / Reliability improvements">​</a></h2>
<ul>
<li>
<p><strong>Memory Leak Fixes</strong></p>
<ul>
<li>Fix: prevent httpx DeprecationWarning memory leak in AsyncHTTPHandler - <a href="https://github.com/BerriAI/litellm/pull/16024" target="_blank" rel="noopener noreferrer">PR #16024</a></li>
<li>Fix: resolve memory accumulation caused by Pydantic 2.11+ deprecation warnings - <a href="https://github.com/BerriAI/litellm/pull/16110" target="_blank" rel="noopener noreferrer">PR #16110</a></li>
<li>Fix(apscheduler): prevent memory leaks from jitter and frequent job intervals - <a href="https://github.com/BerriAI/litellm/pull/15846" target="_blank" rel="noopener noreferrer">PR #15846</a></li>
</ul>
</li>
<li>
<p><strong>Configuration</strong></p>
<ul>
<li>Remove minimum validation for cache control injection index - <a href="https://github.com/BerriAI/litellm/pull/16149" target="_blank" rel="noopener noreferrer">PR #16149</a></li>
<li>Fix prompt_caching.md: wrong prompt_tokens definition - <a href="https://github.com/BerriAI/litellm/pull/16044" target="_blank" rel="noopener noreferrer">PR #16044</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates">Documentation Updates<a href="https://docs.litellm.ai/release_notes/v1-79-1#documentation-updates" class="hash-link" aria-label="Direct link to Documentation Updates" title="Direct link to Documentation Updates">​</a></h2>
<ul>
<li>
<p><strong>Provider Documentation</strong></p>
<ul>
<li>Use custom-llm-provider header in examples - <a href="https://github.com/BerriAI/litellm/pull/16055" target="_blank" rel="noopener noreferrer">PR #16055</a></li>
<li>Litellm docs readme fixes - <a href="https://github.com/BerriAI/litellm/pull/16107" target="_blank" rel="noopener noreferrer">PR #16107</a></li>
<li>Readme fixes add supported providers - <a href="https://github.com/BerriAI/litellm/pull/16109" target="_blank" rel="noopener noreferrer">PR #16109</a></li>
</ul>
</li>
<li>
<p><strong>Model References</strong></p>
<ul>
<li>Add supports vision field to qwen-vl models in model_prices_and_context_window.json - <a href="https://github.com/BerriAI/litellm/pull/16106" target="_blank" rel="noopener noreferrer">PR #16106</a></li>
</ul>
</li>
<li>
<p><strong>General Documentation</strong></p>
<ul>
<li>1-79-0 docs - <a href="https://github.com/BerriAI/litellm/pull/15936" target="_blank" rel="noopener noreferrer">PR #15936</a></li>
<li>Add minimum resource requirement for production - <a href="https://github.com/BerriAI/litellm/pull/16146" target="_blank" rel="noopener noreferrer">PR #16146</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-79-1#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@RobGeada made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15975" target="_blank" rel="noopener noreferrer">PR #15975</a></li>
<li>@shanto12 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15946" target="_blank" rel="noopener noreferrer">PR #15946</a></li>
<li>@dima-hx430 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15976" target="_blank" rel="noopener noreferrer">PR #15976</a></li>
<li>@m-misiura made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15971" target="_blank" rel="noopener noreferrer">PR #15971</a></li>
<li>@ylgibby made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15947" target="_blank" rel="noopener noreferrer">PR #15947</a></li>
<li>@Somtom made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15909" target="_blank" rel="noopener noreferrer">PR #15909</a></li>
<li>@rodolfo-nobrega made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16023" target="_blank" rel="noopener noreferrer">PR #16023</a></li>
<li>@bernata made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15997" target="_blank" rel="noopener noreferrer">PR #15997</a></li>
<li>@AlbertDeFusco made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15881" target="_blank" rel="noopener noreferrer">PR #15881</a></li>
<li>@komarovd95 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15789" target="_blank" rel="noopener noreferrer">PR #15789</a></li>
<li>@langpingxue made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15635" target="_blank" rel="noopener noreferrer">PR #15635</a></li>
<li>@OrionCodeDev made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16070" target="_blank" rel="noopener noreferrer">PR #16070</a></li>
<li>@sbinnee made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16078" target="_blank" rel="noopener noreferrer">PR #16078</a></li>
<li>@JetoPistola made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16106" target="_blank" rel="noopener noreferrer">PR #16106</a></li>
<li>@gvioss made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16093" target="_blank" rel="noopener noreferrer">PR #16093</a></li>
<li>@pale-aura made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16084" target="_blank" rel="noopener noreferrer">PR #16084</a></li>
<li>@tanvithakur94 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16041" target="_blank" rel="noopener noreferrer">PR #16041</a></li>
<li>@li-boxuan made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16044" target="_blank" rel="noopener noreferrer">PR #16044</a></li>
<li>@1stprinciple made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15938" target="_blank" rel="noopener noreferrer">PR #15938</a></li>
<li>@raghav-stripe made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16137" target="_blank" rel="noopener noreferrer">PR #16137</a></li>
<li>@steve-gore-snapdocs made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/16149" target="_blank" rel="noopener noreferrer">PR #16149</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog">Full Changelog<a href="https://docs.litellm.ai/release_notes/v1-79-1#full-changelog" class="hash-link" aria-label="Direct link to Full Changelog" title="Direct link to Full Changelog">​</a></h2>
<p><strong><a href="https://github.com/BerriAI/litellm/compare/v1.79.0-stable...v1.80.0-stable" target="_blank" rel="noopener noreferrer">View complete changelog on GitHub</a></strong></p>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[v1.79.0-stable - Search APIs]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-79-0</id>
        <link href="https://docs.litellm.ai/release_notes/v1-79-0"/>
        <updated>2025-10-26T10:00:00.000Z</updated>
        <summary type="html"><![CDATA[Deploy this version]]></summary>
        <content type="html"><![CDATA[<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-79-0#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">docker.litellm.ai/berriai/litellm:v1.79.0-stable</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.79.0</span><br></span></code></pre></div></div></div></div></div>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="major-changes">Major Changes<a href="https://docs.litellm.ai/release_notes/v1-79-0#major-changes" class="hash-link" aria-label="Direct link to Major Changes" title="Direct link to Major Changes">​</a></h2>
<ul>
<li><strong>Cohere models will now be routed to Cohere v2 API by default</strong> - <a href="https://github.com/BerriAI/litellm/pull/15722" target="_blank" rel="noopener noreferrer">PR #15722</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-79-0#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<ul>
<li><strong>Search APIs</strong> - Native <code>/v1/search</code> endpoint with support for Perplexity, Tavily, Parallel AI, Exa AI, DataforSEO, and Google PSE with cost tracking</li>
<li><strong>Vector Stores</strong> - Vertex AI Search API integration as vector store through LiteLLM with passthrough endpoint support</li>
<li><strong>Guardrails Expansion</strong> - Apply guardrails across Responses API, Image Gen, Text completions, Audio transcriptions, Audio Speech, Rerank, and Anthropic Messages API via unified <code>apply_guardrails</code> function</li>
<li><strong>New Guardrail Providers</strong> - Gray Swan, Dynamo AI, IBM Guardrails, Lasso Security v3, and Bedrock Guardrail apply_guardrail endpoint support</li>
<li><strong>Video Generation API</strong> - Native support for OpenAI Sora-2 and Azure Sora-2 (Pro, Pro-High-Res) with cost tracking and logging support</li>
<li><strong>Azure AI Speech (TTS)</strong> - Native Azure AI Speech integration with cost tracking for standard and HD voices</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-79-0#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support">New Model Support<a href="https://docs.litellm.ai/release_notes/v1-79-0#new-model-support" class="hash-link" aria-label="Direct link to New Model Support" title="Direct link to New Model Support">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th><th>Features</th></tr></thead><tbody><tr><td>Bedrock</td><td><code>anthropic.claude-3-7-sonnet-20240620-v1:0</code></td><td>200K</td><td>$3.60</td><td>$18.00</td><td>Chat, reasoning, vision, function calling, prompt caching, computer use</td></tr><tr><td>Bedrock GovCloud</td><td><code>us-gov-west-1/anthropic.claude-3-7-sonnet-20250219-v1:0</code></td><td>200K</td><td>$3.60</td><td>$18.00</td><td>Chat, reasoning, vision, function calling, prompt caching, computer use</td></tr><tr><td>Vertex AI</td><td><code>mistral-medium-3</code></td><td>128K</td><td>$0.40</td><td>$2.00</td><td>Chat, function calling, tool choice</td></tr><tr><td>Vertex AI</td><td><code>codestral-2</code></td><td>128K</td><td>$0.30</td><td>$0.90</td><td>Chat, function calling, tool choice</td></tr><tr><td>Bedrock</td><td><code>amazon.titan-image-generator-v1</code></td><td>-</td><td>-</td><td>-</td><td>Image generation - $0.008/image, $0.01/premium image</td></tr><tr><td>Bedrock</td><td><code>amazon.titan-image-generator-v2</code></td><td>-</td><td>-</td><td>-</td><td>Image generation - $0.008/image, $0.01/premium image</td></tr><tr><td>OpenAI</td><td><code>sora-2</code></td><td>-</td><td>-</td><td>-</td><td>Video generation - $0.10/video/second</td></tr><tr><td>Azure</td><td><code>sora-2</code></td><td>-</td><td>-</td><td>-</td><td>Video generation - $0.10/video/second</td></tr><tr><td>Azure</td><td><code>sora-2-pro</code></td><td>-</td><td>-</td><td>-</td><td>Video generation - $0.30/video/second</td></tr><tr><td>Azure</td><td><code>sora-2-pro-high-res</code></td><td>-</td><td>-</td><td>-</td><td>Video generation - $0.50/video/second</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-79-0#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Fix cache_control incorrectly applied to all content items instead of last item only - <a href="https://github.com/BerriAI/litellm/pull/15699" target="_blank" rel="noopener noreferrer">PR #15699</a></li>
<li>Forward anthropic-beta headers to Bedrock, VertexAI - <a href="https://github.com/BerriAI/litellm/pull/15700" target="_blank" rel="noopener noreferrer">PR #15700</a></li>
<li>Change max_tokens value to match max_output_tokens for claude sonnet - <a href="https://github.com/BerriAI/litellm/pull/15715" target="_blank" rel="noopener noreferrer">PR #15715</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong></p>
<ul>
<li>Add AWS us-gov-west-1 Claude 3.7 Sonnet costs - <a href="https://github.com/BerriAI/litellm/pull/15775" target="_blank" rel="noopener noreferrer">PR #15775</a></li>
<li>Fix the date for sonnet 3.7 in govcloud - <a href="https://github.com/BerriAI/litellm/pull/15800" target="_blank" rel="noopener noreferrer">PR #15800</a></li>
<li>Use proper bedrock model name in health check - <a href="https://github.com/BerriAI/litellm/pull/15808" target="_blank" rel="noopener noreferrer">PR #15808</a></li>
<li>Support for embeddings_by_type Response Format in Bedrock Cohere Embed v1 - <a href="https://github.com/BerriAI/litellm/pull/15707" target="_blank" rel="noopener noreferrer">PR #15707</a></li>
<li>Add titan image generations with cost tracking - <a href="https://github.com/BerriAI/litellm/pull/15916" target="_blank" rel="noopener noreferrer">PR #15916</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Gemini</a></strong></p>
<ul>
<li>Add imageConfig parameter for gemini-2.5-flash-image - <a href="https://github.com/BerriAI/litellm/pull/15530" target="_blank" rel="noopener noreferrer">PR #15530</a></li>
<li>Replace deprecated gemini-1.5-pro-preview-0514 - <a href="https://github.com/BerriAI/litellm/pull/15852" target="_blank" rel="noopener noreferrer">PR #15852</a></li>
<li>Update vertex ai gemini costs - <a href="https://github.com/BerriAI/litellm/pull/15911" target="_blank" rel="noopener noreferrer">PR #15911</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/ollama">Ollama</a></strong></p>
<ul>
<li>Set 'think' to False when reasoning effort is minimal/none/disable - <a href="https://github.com/BerriAI/litellm/pull/15763" target="_blank" rel="noopener noreferrer">PR #15763</a></li>
<li>Handle parsing ollama chunk error - <a href="https://github.com/BerriAI/litellm/pull/15717" target="_blank" rel="noopener noreferrer">PR #15717</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vertex">Vertex AI</a></strong></p>
<ul>
<li>Add mistral medium 3 and Codestral 2 on vertex - <a href="https://github.com/BerriAI/litellm/pull/15887" target="_blank" rel="noopener noreferrer">PR #15887</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/databricks">Databricks</a></strong></p>
<ul>
<li>Allow prompt caching to be used for Anthropic Claude on Databricks - <a href="https://github.com/BerriAI/litellm/pull/15801" target="_blank" rel="noopener noreferrer">PR #15801</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/azure">Azure</a></strong></p>
<ul>
<li>Add Azure AVA TTS integration - <a href="https://github.com/BerriAI/litellm/pull/15749" target="_blank" rel="noopener noreferrer">PR #15749</a></li>
<li>Add Azure AVA (Speech AI) Cost Tracking - <a href="https://github.com/BerriAI/litellm/pull/15754" target="_blank" rel="noopener noreferrer">PR #15754</a></li>
<li>Azure AI Speech - Ensure <code>voice</code> is mapped from request body to SSML body, allow sending <code>role</code> and <code>style</code> - <a href="https://github.com/BerriAI/litellm/pull/15810" target="_blank" rel="noopener noreferrer">PR #15810</a></li>
<li>Add Azure support for video generation functionality (Sora-2) - <a href="https://github.com/BerriAI/litellm/pull/15901" target="_blank" rel="noopener noreferrer">PR #15901</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI</a></strong></p>
<ul>
<li>OpenAI videos refactoring - <a href="https://github.com/BerriAI/litellm/pull/15900" target="_blank" rel="noopener noreferrer">PR #15900</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Read from custom-llm-provider header - <a href="https://github.com/BerriAI/litellm/pull/15528" target="_blank" rel="noopener noreferrer">PR #15528</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-79-0#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-79-0#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/response_api">Responses API</a></strong></p>
<ul>
<li>Add gpt 4.1 pricing for response endpoint - <a href="https://github.com/BerriAI/litellm/pull/15593" target="_blank" rel="noopener noreferrer">PR #15593</a></li>
<li>Fix Incorrect status value in responses api with gemini - <a href="https://github.com/BerriAI/litellm/pull/15753" target="_blank" rel="noopener noreferrer">PR #15753</a></li>
<li>Simplify reasoning item handling for gpt-5-codex - <a href="https://github.com/BerriAI/litellm/pull/15815" target="_blank" rel="noopener noreferrer">PR #15815</a></li>
<li>ErrorEvent ValidationError when OpenAI Responses API returns nested error structure - <a href="https://github.com/BerriAI/litellm/pull/15804" target="_blank" rel="noopener noreferrer">PR #15804</a></li>
<li>Fix reasoning item ID auto-generation causing encrypted content verification errors - <a href="https://github.com/BerriAI/litellm/pull/15782" target="_blank" rel="noopener noreferrer">PR #15782</a></li>
<li>Support tags in metadata - <a href="https://github.com/BerriAI/litellm/pull/15867" target="_blank" rel="noopener noreferrer">PR #15867</a></li>
<li>Security: prevent User A from retrieving User B's response, if response.id is leaked - <a href="https://github.com/BerriAI/litellm/pull/15757" target="_blank" rel="noopener noreferrer">PR #15757</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/batch_api">Batch API</a></strong></p>
<ul>
<li>Add pre and post call for list batches - <a href="https://github.com/BerriAI/litellm/pull/15673" target="_blank" rel="noopener noreferrer">PR #15673</a></li>
<li>Add function responsible to call precall - <a href="https://github.com/BerriAI/litellm/pull/15636" target="_blank" rel="noopener noreferrer">PR #15636</a></li>
<li>Fix "User default_user_id does not have access to the object" when object not in db - <a href="https://github.com/BerriAI/litellm/pull/15873" target="_blank" rel="noopener noreferrer">PR #15873</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/ocr">OCR API</a></strong></p>
<ul>
<li>Add Azure AI - OCR to docs - <a href="https://github.com/BerriAI/litellm/pull/15768" target="_blank" rel="noopener noreferrer">PR #15768</a></li>
<li>Add mode + Health check support for OCR models - <a href="https://github.com/BerriAI/litellm/pull/15767" target="_blank" rel="noopener noreferrer">PR #15767</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/search_api">Search API</a></strong></p>
<ul>
<li>Add def search() APIs for Web Search - Perplexity API - <a href="https://github.com/BerriAI/litellm/pull/15769" target="_blank" rel="noopener noreferrer">PR #15769</a></li>
<li>Add Tavily Search API - <a href="https://github.com/BerriAI/litellm/pull/15770" target="_blank" rel="noopener noreferrer">PR #15770</a></li>
<li>Add Parallel AI - Search API - <a href="https://github.com/BerriAI/litellm/pull/15772" target="_blank" rel="noopener noreferrer">PR #15772</a></li>
<li>Add EXA AI Search API to LiteLLM - <a href="https://github.com/BerriAI/litellm/pull/15774" target="_blank" rel="noopener noreferrer">PR #15774</a></li>
<li>Add /search endpoint on LiteLLM Gateway - <a href="https://github.com/BerriAI/litellm/pull/15780" target="_blank" rel="noopener noreferrer">PR #15780</a></li>
<li>Add DataforSEO Search API - <a href="https://github.com/BerriAI/litellm/pull/15817" target="_blank" rel="noopener noreferrer">PR #15817</a></li>
<li>Add Google PSE Search Provider - <a href="https://github.com/BerriAI/litellm/pull/15816" target="_blank" rel="noopener noreferrer">PR #15816</a></li>
<li>Add cost tracking for Search API requests - Google PSE, Tavily, Parallel AI, Exa AI - <a href="https://github.com/BerriAI/litellm/pull/15821" target="_blank" rel="noopener noreferrer">PR #15821</a></li>
<li>Backend: Allow storing configured Search APIs in DB - <a href="https://github.com/BerriAI/litellm/pull/15862" target="_blank" rel="noopener noreferrer">PR #15862</a></li>
<li>Exa Search API - ensure request params are sent to Exa AI - <a href="https://github.com/BerriAI/litellm/pull/15855" target="_blank" rel="noopener noreferrer">PR #15855</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/vector_stores">Vector Stores</a></strong></p>
<ul>
<li>Support Vertex AI Search API as vector store through LiteLLM - <a href="https://github.com/BerriAI/litellm/pull/15781" target="_blank" rel="noopener noreferrer">PR #15781</a></li>
<li>Azure AI - Search Vector Stores - <a href="https://github.com/BerriAI/litellm/pull/15873" target="_blank" rel="noopener noreferrer">PR #15873</a></li>
<li>VertexAI Search Vector Store - Passthrough endpoint support + Vector store search Cost tracking support - <a href="https://github.com/BerriAI/litellm/pull/15824" target="_blank" rel="noopener noreferrer">PR #15824</a></li>
<li>Don't raise error if managed object is not found - <a href="https://github.com/BerriAI/litellm/pull/15873" target="_blank" rel="noopener noreferrer">PR #15873</a></li>
<li>Show config.yaml vector stores on UI - <a href="https://github.com/BerriAI/litellm/pull/15873" target="_blank" rel="noopener noreferrer">PR #15873</a></li>
<li>Cost tracking for search spend - <a href="https://github.com/BerriAI/litellm/pull/15859" target="_blank" rel="noopener noreferrer">PR #15859</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/image_generation">Images API</a></strong></p>
<ul>
<li>Pass user-defined headers and extra_headers to image-edit calls - <a href="https://github.com/BerriAI/litellm/pull/15811" target="_blank" rel="noopener noreferrer">PR #15811</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/video_generation">Video Generation API</a></strong></p>
<ul>
<li>Add Azure support for video generation functionality (Sora-2, Sora-2-Pro, Sora-2-Pro-High-Res) - <a href="https://github.com/BerriAI/litellm/pull/15901" target="_blank" rel="noopener noreferrer">PR #15901</a></li>
<li>OpenAI video generation refactoring (Sora-2) - <a href="https://github.com/BerriAI/litellm/pull/15900" target="_blank" rel="noopener noreferrer">PR #15900</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/bedrock_invoke">Bedrock /invoke</a></strong></p>
<ul>
<li>Fix: Hooks broken on /bedrock passthrough due to missing metadata - <a href="https://github.com/BerriAI/litellm/pull/15849" target="_blank" rel="noopener noreferrer">PR #15849</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/realtime_api">Realtime API</a></strong></p>
<ul>
<li>Fix: OpenAI Realtime API integration fails due to websockets.exceptions.PayloadTooBig error - <a href="https://github.com/BerriAI/litellm/pull/15751" target="_blank" rel="noopener noreferrer">PR #15751</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-79-0#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-79-0#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong>Passthrough</strong></p>
<ul>
<li>Set auth on passthrough endpoints, on the UI - <a href="https://github.com/BerriAI/litellm/pull/15778" target="_blank" rel="noopener noreferrer">PR #15778</a></li>
<li>Fix pass-through endpoint budget enforcement bug - <a href="https://github.com/BerriAI/litellm/pull/15805" target="_blank" rel="noopener noreferrer">PR #15805</a></li>
</ul>
</li>
<li>
<p><strong>Organizations</strong></p>
<ul>
<li>Allow org admins to create teams on UI - <a href="https://github.com/BerriAI/litellm/pull/15924" target="_blank" rel="noopener noreferrer">PR #15924</a></li>
</ul>
</li>
<li>
<p><strong>Search Tools</strong></p>
<ul>
<li>UI - Search Tools, allow adding search tools on UI + testing search - <a href="https://github.com/BerriAI/litellm/pull/15871" target="_blank" rel="noopener noreferrer">PR #15871</a></li>
<li>UI - Add logos for search providers - <a href="https://github.com/BerriAI/litellm/pull/15872" target="_blank" rel="noopener noreferrer">PR #15872</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Fix routing for custom server root path - <a href="https://github.com/BerriAI/litellm/pull/15701" target="_blank" rel="noopener noreferrer">PR #15701</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="logging--guardrail--prompt-management-integrations">Logging / Guardrail / Prompt Management Integrations<a href="https://docs.litellm.ai/release_notes/v1-79-0#logging--guardrail--prompt-management-integrations" class="hash-link" aria-label="Direct link to Logging / Guardrail / Prompt Management Integrations" title="Direct link to Logging / Guardrail / Prompt Management Integrations">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-3">Features<a href="https://docs.litellm.ai/release_notes/v1-79-0#features-3" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#opentelemetry">OpenTelemetry</a></strong></p>
<ul>
<li>Fix OpenTelemetry Logging functionality - <a href="https://github.com/BerriAI/litellm/pull/15645" target="_blank" rel="noopener noreferrer">PR #15645</a></li>
<li>Fix issue where headers were not being split correctly - <a href="https://github.com/BerriAI/litellm/pull/15916" target="_blank" rel="noopener noreferrer">PR #15916</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#sentry">Sentry</a></strong></p>
<ul>
<li>Add SENTRY_ENVIRONMENT configuration for Sentry integration - <a href="https://github.com/BerriAI/litellm/pull/15760" target="_blank" rel="noopener noreferrer">PR #15760</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#helicone">Helicone</a></strong></p>
<ul>
<li>Fix JSON serialization error in Helicone logging by removing OpenTelemetry span from metadata - <a href="https://github.com/BerriAI/litellm/pull/15728" target="_blank" rel="noopener noreferrer">PR #15728</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/logging#mlflow">MLFlow</a></strong></p>
<ul>
<li>Fix MLFlow tags - split request_tags into (key, val) if request_tag has colon - <a href="https://github.com/BerriAI/litellm/pull/15914" target="_blank" rel="noopener noreferrer">PR #15914</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Rename configured_cold_storage_logger to cold_storage_custom_logger - <a href="https://github.com/BerriAI/litellm/pull/15798" target="_blank" rel="noopener noreferrer">PR #15798</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails">Guardrails<a href="https://docs.litellm.ai/release_notes/v1-79-0#guardrails" class="hash-link" aria-label="Direct link to Guardrails" title="Direct link to Guardrails">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">Gray Swan</a></strong></p>
<ul>
<li>Add GraySwan Guardrails support - <a href="https://github.com/BerriAI/litellm/pull/15756" target="_blank" rel="noopener noreferrer">PR #15756</a></li>
<li>Rename GraySwan to Gray Swan - <a href="https://github.com/BerriAI/litellm/pull/15771" target="_blank" rel="noopener noreferrer">PR #15771</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">Dynamo AI</a></strong></p>
<ul>
<li>New Guardrail - Dynamo AI Guardrail - <a href="https://github.com/BerriAI/litellm/pull/15920" target="_blank" rel="noopener noreferrer">PR #15920</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">IBM Guardrails</a></strong></p>
<ul>
<li>IBM Guardrails integration - <a href="https://github.com/BerriAI/litellm/pull/15924" target="_blank" rel="noopener noreferrer">PR #15924</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">Lasso Security</a></strong></p>
<ul>
<li>Add v3 API Support - <a href="https://github.com/BerriAI/litellm/pull/12452" target="_blank" rel="noopener noreferrer">PR #12452</a></li>
<li>Fixed lasso import config, redis cluster hash tags for test keys - <a href="https://github.com/BerriAI/litellm/pull/15917" target="_blank" rel="noopener noreferrer">PR #15917</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">Bedrock Guardrails</a></strong></p>
<ul>
<li>Implement Bedrock Guardrail apply_guardrail endpoint support - <a href="https://github.com/BerriAI/litellm/pull/15892" target="_blank" rel="noopener noreferrer">PR #15892</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Guardrails - Responses API, Image Gen, Text completions, Audio transcriptions, Audio Speech, Rerank, Anthropic Messages API support via the unified <code>apply_guardrails</code> function - <a href="https://github.com/BerriAI/litellm/pull/15706" target="_blank" rel="noopener noreferrer">PR #15706</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-79-0#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li><strong>Rate Limiting</strong>
<ul>
<li>Support absolute RPM/TPM in priority_reservation - <a href="https://github.com/BerriAI/litellm/pull/15813" target="_blank" rel="noopener noreferrer">PR #15813</a></li>
<li>Org level tpm/rpm limits + Team tpm/rpm validation when assigned to org - <a href="https://github.com/BerriAI/litellm/pull/15549" target="_blank" rel="noopener noreferrer">PR #15549</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-gateway">MCP Gateway<a href="https://docs.litellm.ai/release_notes/v1-79-0#mcp-gateway" class="hash-link" aria-label="Direct link to MCP Gateway" title="Direct link to MCP Gateway">​</a></h2>
<ul>
<li><strong>OAuth</strong>
<ul>
<li>Auth Header Fix for MCP Tool Call - <a href="https://github.com/BerriAI/litellm/pull/15736" target="_blank" rel="noopener noreferrer">PR #15736</a></li>
<li>Add response_type + PKCE parameters to OAuth authorization endpoint - <a href="https://github.com/BerriAI/litellm/pull/15720" target="_blank" rel="noopener noreferrer">PR #15720</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements">Performance / Loadbalancing / Reliability improvements<a href="https://docs.litellm.ai/release_notes/v1-79-0#performance--loadbalancing--reliability-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements" title="Direct link to Performance / Loadbalancing / Reliability improvements">​</a></h2>
<ul>
<li>
<p><strong>Database</strong></p>
<ul>
<li>Minimize the occurrence of deadlocks - <a href="https://github.com/BerriAI/litellm/pull/15281" target="_blank" rel="noopener noreferrer">PR #15281</a></li>
</ul>
</li>
<li>
<p><strong>Redis</strong></p>
<ul>
<li>Apply max_connections configuration to Redis async client - <a href="https://github.com/BerriAI/litellm/pull/15797" target="_blank" rel="noopener noreferrer">PR #15797</a></li>
</ul>
</li>
<li>
<p><strong>Caching</strong></p>
<ul>
<li>Add documentation for <code>enable_caching_on_provider_specific_optional_params</code> setting - <a href="https://github.com/BerriAI/litellm/pull/15885" target="_blank" rel="noopener noreferrer">PR #15885</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates">Documentation Updates<a href="https://docs.litellm.ai/release_notes/v1-79-0#documentation-updates" class="hash-link" aria-label="Direct link to Documentation Updates" title="Direct link to Documentation Updates">​</a></h2>
<ul>
<li><strong>Provider Documentation</strong>
<ul>
<li>Update worker recommendation - <a href="https://github.com/BerriAI/litellm/pull/15702" target="_blank" rel="noopener noreferrer">PR #15702</a></li>
<li>Fix the wrong request body in json mode doc - <a href="https://github.com/BerriAI/litellm/pull/15729" target="_blank" rel="noopener noreferrer">PR #15729</a></li>
<li>Add details in docs - <a href="https://github.com/BerriAI/litellm/pull/15721" target="_blank" rel="noopener noreferrer">PR #15721</a></li>
<li>Add responses api on openai docs - <a href="https://github.com/BerriAI/litellm/pull/15866" target="_blank" rel="noopener noreferrer">PR #15866</a></li>
<li>Add OpenAI responses api - <a href="https://github.com/BerriAI/litellm/pull/15868" target="_blank" rel="noopener noreferrer">PR #15868</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-79-0#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@tlecomte made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15528" target="_blank" rel="noopener noreferrer">PR #15528</a></li>
<li>@tomhaynes made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15645" target="_blank" rel="noopener noreferrer">PR #15645</a></li>
<li>@talalryz made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15720" target="_blank" rel="noopener noreferrer">PR #15720</a></li>
<li>@1vinodsingh1 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15736" target="_blank" rel="noopener noreferrer">PR #15736</a></li>
<li>@nuernber made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15775" target="_blank" rel="noopener noreferrer">PR #15775</a></li>
<li>@Thomas-Mildner made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15760" target="_blank" rel="noopener noreferrer">PR #15760</a></li>
<li>@javiergarciapleo made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15721" target="_blank" rel="noopener noreferrer">PR #15721</a></li>
<li>@lshgdut made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15717" target="_blank" rel="noopener noreferrer">PR #15717</a></li>
<li>@kk-wangjifeng made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15530" target="_blank" rel="noopener noreferrer">PR #15530</a></li>
<li>@anthonyivn2 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15801" target="_blank" rel="noopener noreferrer">PR #15801</a></li>
<li>@romanglo made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15707" target="_blank" rel="noopener noreferrer">PR #15707</a></li>
<li>@mythral made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15859" target="_blank" rel="noopener noreferrer">PR #15859</a></li>
<li>@mubashirosmani made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15866" target="_blank" rel="noopener noreferrer">PR #15866</a></li>
<li>@CAFxX made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15281" target="_blank" rel="noopener noreferrer">PR #15281</a></li>
<li>@reflection made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15914" target="_blank" rel="noopener noreferrer">PR #15914</a></li>
<li>@shadielfares made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15917" target="_blank" rel="noopener noreferrer">PR #15917</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="pr-count-summary">PR Count Summary<a href="https://docs.litellm.ai/release_notes/v1-79-0#pr-count-summary" class="hash-link" aria-label="Direct link to PR Count Summary" title="Direct link to PR Count Summary">​</a></h2>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="10262025">10/26/2025<a href="https://docs.litellm.ai/release_notes/v1-79-0#10262025" class="hash-link" aria-label="Direct link to 10/26/2025" title="Direct link to 10/26/2025">​</a></h3>
<ul>
<li>New Models / Updated Models: 20</li>
<li>LLM API Endpoints: 29</li>
<li>Management Endpoints / UI: 5</li>
<li>Logging / Guardrail / Prompt Management Integrations: 10</li>
<li>Spend Tracking, Budgets and Rate Limiting: 2</li>
<li>MCP Gateway: 2</li>
<li>Performance / Loadbalancing / Reliability improvements: 3</li>
<li>Documentation Updates: 5</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog">Full Changelog<a href="https://docs.litellm.ai/release_notes/v1-79-0#full-changelog" class="hash-link" aria-label="Direct link to Full Changelog" title="Direct link to Full Changelog">​</a></h2>
<p><strong><a href="https://github.com/BerriAI/litellm/compare/v1.78.5-stable...v1.79.0-stable" target="_blank" rel="noopener noreferrer">View complete changelog on GitHub</a></strong></p>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[v1.78.5-stable - Native OCR Support]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-78-5</id>
        <link href="https://docs.litellm.ai/release_notes/v1-78-5"/>
        <updated>2025-10-18T10:00:00.000Z</updated>
        <summary type="html"><![CDATA[Deploy this version]]></summary>
        <content type="html"><![CDATA[<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-78-5#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">docker.litellm.ai/berriai/litellm:v1.78.5-stable</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.78.5</span><br></span></code></pre></div></div></div></div></div>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-78-5#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<ul>
<li><strong>Native OCR Endpoints</strong> - Native <code>/v1/ocr</code> endpoint support with cost tracking for Mistral OCR and Azure AI OCR</li>
<li><strong>Global Vendor Discounts</strong> - Specify global vendor discount percentages for accurate cost tracking and reporting</li>
<li><strong>Team Spending Reports</strong> - Team admins can now export detailed spending reports for their teams</li>
<li><strong>Claude Haiku 4.5</strong> - Day 0 support for Claude Haiku 4.5 across Bedrock, Vertex AI, and OpenRouter with 200K context window</li>
<li><strong>GPT-5-Codex</strong> - Support for GPT-5-Codex via Responses API on OpenAI and Azure</li>
<li><strong>Performance Improvements</strong> - Major router optimizations: O(1) model lookups, 10-100x faster shallow copy, 30-40% faster timing calls, and O(n) to O(1) hash generation</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-78-5#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support">New Model Support<a href="https://docs.litellm.ai/release_notes/v1-78-5#new-model-support" class="hash-link" aria-label="Direct link to New Model Support" title="Direct link to New Model Support">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th><th>Features</th></tr></thead><tbody><tr><td>Anthropic</td><td><code>claude-haiku-4-5</code></td><td>200K</td><td>$1.00</td><td>$5.00</td><td>Chat, reasoning, vision, function calling, prompt caching, computer use</td></tr><tr><td>Anthropic</td><td><code>claude-haiku-4-5-20251001</code></td><td>200K</td><td>$1.00</td><td>$5.00</td><td>Chat, reasoning, vision, function calling, prompt caching, computer use</td></tr><tr><td>Bedrock</td><td><code>anthropic.claude-haiku-4-5-20251001-v1:0</code></td><td>200K</td><td>$1.00</td><td>$5.00</td><td>Chat, reasoning, vision, function calling, prompt caching</td></tr><tr><td>Bedrock</td><td><code>global.anthropic.claude-haiku-4-5-20251001-v1:0</code></td><td>200K</td><td>$1.00</td><td>$5.00</td><td>Chat, reasoning, vision, function calling, prompt caching</td></tr><tr><td>Bedrock</td><td><code>jp.anthropic.claude-haiku-4-5-20251001-v1:0</code></td><td>200K</td><td>$1.10</td><td>$5.50</td><td>Chat, reasoning, vision, function calling, prompt caching (JP Cross-Region)</td></tr><tr><td>Bedrock</td><td><code>us.anthropic.claude-haiku-4-5-20251001-v1:0</code></td><td>200K</td><td>$1.10</td><td>$5.50</td><td>Chat, reasoning, vision, function calling, prompt caching (US region)</td></tr><tr><td>Bedrock</td><td><code>eu.anthropic.claude-haiku-4-5-20251001-v1:0</code></td><td>200K</td><td>$1.10</td><td>$5.50</td><td>Chat, reasoning, vision, function calling, prompt caching (EU region)</td></tr><tr><td>Bedrock</td><td><code>apac.anthropic.claude-haiku-4-5-20251001-v1:0</code></td><td>200K</td><td>$1.10</td><td>$5.50</td><td>Chat, reasoning, vision, function calling, prompt caching (APAC region)</td></tr><tr><td>Bedrock</td><td><code>au.anthropic.claude-haiku-4-5-20251001-v1:0</code></td><td>200K</td><td>$1.10</td><td>$5.50</td><td>Chat, reasoning, vision, function calling, prompt caching (AU region)</td></tr><tr><td>Vertex AI</td><td><code>vertex_ai/claude-haiku-4-5@20251001</code></td><td>200K</td><td>$1.00</td><td>$5.00</td><td>Chat, reasoning, vision, function calling, prompt caching</td></tr><tr><td>OpenAI</td><td><code>gpt-5</code></td><td>272K</td><td>$1.25</td><td>$10.00</td><td>Chat, responses API, reasoning, vision, function calling, prompt caching</td></tr><tr><td>OpenAI</td><td><code>gpt-5-codex</code></td><td>272K</td><td>$1.25</td><td>$10.00</td><td>Responses API mode</td></tr><tr><td>Azure</td><td><code>azure/gpt-5-codex</code></td><td>272K</td><td>$1.25</td><td>$10.00</td><td>Responses API mode</td></tr><tr><td>Gemini</td><td><code>gemini-2.5-flash-image</code></td><td>32K</td><td>$0.30</td><td>$2.50</td><td>Image generation (GA - Nano Banana) - $0.039/image</td></tr><tr><td>ZhipuAI</td><td><code>glm-4.6</code></td><td>-</td><td>-</td><td>-</td><td>Chat completions</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-78-5#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI</a></strong></p>
<ul>
<li>GPT-5 return reasoning content via /chat/completions + GPT-5-Codex working on Claude Code - <a href="https://github.com/BerriAI/litellm/pull/15441" target="_blank" rel="noopener noreferrer">PR #15441</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Reduce claude-4-sonnet max_output_tokens to 64k - <a href="https://github.com/BerriAI/litellm/pull/15409" target="_blank" rel="noopener noreferrer">PR #15409</a></li>
<li>Added claude-haiku-4.5 - <a href="https://github.com/BerriAI/litellm/pull/15579" target="_blank" rel="noopener noreferrer">PR #15579</a></li>
<li>Add support for thinking blocks and redacted thinking blocks in Anthropic v1/messages API - <a href="https://github.com/BerriAI/litellm/pull/15501" target="_blank" rel="noopener noreferrer">PR #15501</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong></p>
<ul>
<li>Add anthropic.claude-haiku-4-5-20251001-v1:0 on Bedrock, VertexAI - <a href="https://github.com/BerriAI/litellm/pull/15581" target="_blank" rel="noopener noreferrer">PR #15581</a></li>
<li>Add Claude Haiku 4.5 support for Bedrock global and US regions - <a href="https://github.com/BerriAI/litellm/pull/15650" target="_blank" rel="noopener noreferrer">PR #15650</a></li>
<li>Add Claude Haiku 4.5 support for Bedrock Other regions - <a href="https://github.com/BerriAI/litellm/pull/15653" target="_blank" rel="noopener noreferrer">PR #15653</a></li>
<li>Add JP Cross-Region Inference jp.anthropic.claude-haiku-4-5-20251001 - <a href="https://github.com/BerriAI/litellm/pull/15598" target="_blank" rel="noopener noreferrer">PR #15598</a></li>
<li>Fix: bedrock-pricing-geo-inregion-cross-region / add Global Cross-Region Inference - <a href="https://github.com/BerriAI/litellm/pull/15685" target="_blank" rel="noopener noreferrer">PR #15685</a></li>
<li>Fix: Support us-gov prefix for AWS GovCloud Bedrock models - <a href="https://github.com/BerriAI/litellm/pull/15626" target="_blank" rel="noopener noreferrer">PR #15626</a></li>
<li>Fix GPT-OSS in Bedrock now supports streaming. Revert fake streaming - <a href="https://github.com/BerriAI/litellm/pull/15668" target="_blank" rel="noopener noreferrer">PR #15668</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Gemini</a></strong></p>
<ul>
<li>Feat(pricing): Add Gemini 2.5 Flash Image (Nano Banana) in GA - <a href="https://github.com/BerriAI/litellm/pull/15557" target="_blank" rel="noopener noreferrer">PR #15557</a></li>
<li>Fix: Gemini 2.5 Flash Image should not have supports_web_search=true - <a href="https://github.com/BerriAI/litellm/pull/15642" target="_blank" rel="noopener noreferrer">PR #15642</a></li>
<li>Remove penalty params as supported params for gemini preview model - <a href="https://github.com/BerriAI/litellm/pull/15503" target="_blank" rel="noopener noreferrer">PR #15503</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/ollama">Ollama</a></strong></p>
<ul>
<li>Fix(ollama/chat): correctly map reasoning_effort to think in requests - <a href="https://github.com/BerriAI/litellm/pull/15465" target="_blank" rel="noopener noreferrer">PR #15465</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openrouter">OpenRouter</a></strong></p>
<ul>
<li>Add anthropic/claude-sonnet-4.5 to OpenRouter cost map - <a href="https://github.com/BerriAI/litellm/pull/15472" target="_blank" rel="noopener noreferrer">PR #15472</a></li>
<li>Prompt caching for anthropic models with OpenRouter - <a href="https://github.com/BerriAI/litellm/pull/15535" target="_blank" rel="noopener noreferrer">PR #15535</a></li>
<li>Get completion cost directly from OpenRouter - <a href="https://github.com/BerriAI/litellm/pull/15448" target="_blank" rel="noopener noreferrer">PR #15448</a></li>
<li>Fix OpenRouter Claude Opus 4 model naming - <a href="https://github.com/BerriAI/litellm/pull/15495" target="_blank" rel="noopener noreferrer">PR #15495</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/comet">CometAPI</a></strong></p>
<ul>
<li>Fix(cometapi): improve CometAPI provider support (embeddings, image generation, docs) - <a href="https://github.com/BerriAI/litellm/pull/15591" target="_blank" rel="noopener noreferrer">PR #15591</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/lemonade">Lemonade</a></strong></p>
<ul>
<li>Adding new models to the lemonade provider - <a href="https://github.com/BerriAI/litellm/pull/15554" target="_blank" rel="noopener noreferrer">PR #15554</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/watsonx">Watson X</a></strong></p>
<ul>
<li>Fix (pricing): Fix pricing for watsonx model family for various models - <a href="https://github.com/BerriAI/litellm/pull/15670" target="_blank" rel="noopener noreferrer">PR #15670</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vercel_ai_gateway">Vercel AI Gateway</a></strong></p>
<ul>
<li>Add glm-4.6 model to pricing configuration - <a href="https://github.com/BerriAI/litellm/pull/15679" target="_blank" rel="noopener noreferrer">PR #15679</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vertex">Vertex AI</a></strong></p>
<ul>
<li>Add Vertex AI Discovery Engine Rerank Support - <a href="https://github.com/BerriAI/litellm/pull/15532" target="_blank" rel="noopener noreferrer">PR #15532</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-78-5#bug-fixes" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong></p>
<ul>
<li>Fix: Pricing for Claude Sonnet 4.5 in US regions is 10x too high - <a href="https://github.com/BerriAI/litellm/pull/15374" target="_blank" rel="noopener noreferrer">PR #15374</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openrouter">OpenRouter</a></strong></p>
<ul>
<li>Change gpt-5-codex support in model_price json - <a href="https://github.com/BerriAI/litellm/pull/15540" target="_blank" rel="noopener noreferrer">PR #15540</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong></p>
<ul>
<li>Fix filtering headers for signature calcs - <a href="https://github.com/BerriAI/litellm/pull/15590" target="_blank" rel="noopener noreferrer">PR #15590</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Add native reasoning and streaming support flag for gpt-5-codex - <a href="https://github.com/BerriAI/litellm/pull/15569" target="_blank" rel="noopener noreferrer">PR #15569</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-78-5#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-78-5#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/response_api">Responses API</a></strong></p>
<ul>
<li>Responses API - enable calling anthropic/gemini models in Responses API streaming in openai ruby sdk + DB - sanity check pending migrations before startup - <a href="https://github.com/BerriAI/litellm/pull/15432" target="_blank" rel="noopener noreferrer">PR #15432</a></li>
<li>Add support for responses mode in health check - <a href="https://github.com/BerriAI/litellm/pull/15658" target="_blank" rel="noopener noreferrer">PR #15658</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/ocr">OCR API</a></strong></p>
<ul>
<li>Feat: Add native litellm.ocr() functions - <a href="https://github.com/BerriAI/litellm/pull/15567" target="_blank" rel="noopener noreferrer">PR #15567</a></li>
<li>Feat: Add /ocr route on LiteLLM AI Gateway - Adds support for native Mistral OCR calling - <a href="https://github.com/BerriAI/litellm/pull/15571" target="_blank" rel="noopener noreferrer">PR #15571</a></li>
<li>Feat: Add Azure AI Mistral OCR Integration - <a href="https://github.com/BerriAI/litellm/pull/15572" target="_blank" rel="noopener noreferrer">PR #15572</a></li>
<li>Feat: Native /ocr endpoint support - <a href="https://github.com/BerriAI/litellm/pull/15573" target="_blank" rel="noopener noreferrer">PR #15573</a></li>
<li>Feat: Add Cost Tracking for /ocr endpoints - <a href="https://github.com/BerriAI/litellm/pull/15678" target="_blank" rel="noopener noreferrer">PR #15678</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">/generateContent</a></strong></p>
<ul>
<li>Fix: GEMINI - CLI - add google_routes to llm_api_routes - <a href="https://github.com/BerriAI/litellm/pull/15500" target="_blank" rel="noopener noreferrer">PR #15500</a></li>
<li>Fix Pydantic validation error for citationMetadata.citationSources in Google GenAI responses - <a href="https://github.com/BerriAI/litellm/pull/15592" target="_blank" rel="noopener noreferrer">PR #15592</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/image_generation">Images API</a></strong></p>
<ul>
<li>Fix: Dall-e-2 for Image Edits API - <a href="https://github.com/BerriAI/litellm/pull/15604" target="_blank" rel="noopener noreferrer">PR #15604</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/pass_through/bedrock">Bedrock Passthrough</a></strong></p>
<ul>
<li>Feat: Allow calling /invoke, /converse routes through AI Gateway + models on config.yaml - <a href="https://github.com/BerriAI/litellm/pull/15618" target="_blank" rel="noopener noreferrer">PR #15618</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs">Bugs<a href="https://docs.litellm.ai/release_notes/v1-78-5#bugs" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>General</strong>
<ul>
<li>Fix: Convert object to a correct type - <a href="https://github.com/BerriAI/litellm/pull/15634" target="_blank" rel="noopener noreferrer">PR #15634</a></li>
<li>Bug Fix: Tags as metadata dicts were raising exceptions - <a href="https://github.com/BerriAI/litellm/pull/15625" target="_blank" rel="noopener noreferrer">PR #15625</a></li>
<li>Add type hint to function_to_dict and fix typo - <a href="https://github.com/BerriAI/litellm/pull/15580" target="_blank" rel="noopener noreferrer">PR #15580</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-78-5#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-78-5#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong>Virtual Keys</strong></p>
<ul>
<li>Docs: Key Rotations - <a href="https://github.com/BerriAI/litellm/pull/15455" target="_blank" rel="noopener noreferrer">PR #15455</a></li>
<li>Fix: UI - Key Max Budget Removal Error Fix - <a href="https://github.com/BerriAI/litellm/pull/15672" target="_blank" rel="noopener noreferrer">PR #15672</a></li>
<li>litellm_Key Settings Max Budget Removal Error Fix - <a href="https://github.com/BerriAI/litellm/pull/15669" target="_blank" rel="noopener noreferrer">PR #15669</a></li>
</ul>
</li>
<li>
<p><strong>Teams</strong></p>
<ul>
<li>Feat: Allow Team Admins to export a report of the team spending - <a href="https://github.com/BerriAI/litellm/pull/15542" target="_blank" rel="noopener noreferrer">PR #15542</a></li>
</ul>
</li>
<li>
<p><strong>Passthrough</strong></p>
<ul>
<li>Feat: Passthrough - allow admin to give access to specific passthrough endpoints - <a href="https://github.com/BerriAI/litellm/pull/15401" target="_blank" rel="noopener noreferrer">PR #15401</a></li>
</ul>
</li>
<li>
<p><strong>SCIM v2</strong></p>
<ul>
<li>Feat(scim_v2.py): if group.id doesn't exist, use external id + Passthrough - ensure updates and deletions persist across instances - <a href="https://github.com/BerriAI/litellm/pull/15276" target="_blank" rel="noopener noreferrer">PR #15276</a></li>
</ul>
</li>
<li>
<p><strong>SSO</strong></p>
<ul>
<li>Feat: UI SSO - Add PKCE for OKTA SSO - <a href="https://github.com/BerriAI/litellm/pull/15608" target="_blank" rel="noopener noreferrer">PR #15608</a></li>
<li>Fix: Separate OAuth M2M authentication from UI SSO + Handle Introspection endpoint for Oauth2 - <a href="https://github.com/BerriAI/litellm/pull/15667" target="_blank" rel="noopener noreferrer">PR #15667</a></li>
<li>Fix/entraid app roles jwt claim clean - <a href="https://github.com/BerriAI/litellm/pull/15583" target="_blank" rel="noopener noreferrer">PR #15583</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="logging--guardrail--prompt-management-integrations">Logging / Guardrail / Prompt Management Integrations<a href="https://docs.litellm.ai/release_notes/v1-78-5#logging--guardrail--prompt-management-integrations" class="hash-link" aria-label="Direct link to Logging / Guardrail / Prompt Management Integrations" title="Direct link to Logging / Guardrail / Prompt Management Integrations">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails">Guardrails<a href="https://docs.litellm.ai/release_notes/v1-78-5#guardrails" class="hash-link" aria-label="Direct link to Guardrails" title="Direct link to Guardrails">​</a></h4>
<ul>
<li>
<p><strong>General</strong></p>
<ul>
<li>Fix apply_guardrail endpoint returning raw string instead of ApplyGuardrailResponse - <a href="https://github.com/BerriAI/litellm/pull/15436" target="_blank" rel="noopener noreferrer">PR #15436</a></li>
<li>Fix: Ensure guardrail memory sync after database updates - <a href="https://github.com/BerriAI/litellm/pull/15633" target="_blank" rel="noopener noreferrer">PR #15633</a></li>
<li>Feat: add guardrail for image generation - <a href="https://github.com/BerriAI/litellm/pull/15619" target="_blank" rel="noopener noreferrer">PR #15619</a></li>
<li>Feat: Add Guardrails for /v1/messages and /v1/responses API - <a href="https://github.com/BerriAI/litellm/pull/15686" target="_blank" rel="noopener noreferrer">PR #15686</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">Pillar Security</a></strong></p>
<ul>
<li>Feature: update pillar security integration to support no persistence mode in litellm proxy - <a href="https://github.com/BerriAI/litellm/pull/15599" target="_blank" rel="noopener noreferrer">PR #15599</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="prompt-management">Prompt Management<a href="https://docs.litellm.ai/release_notes/v1-78-5#prompt-management" class="hash-link" aria-label="Direct link to Prompt Management" title="Direct link to Prompt Management">​</a></h4>
<ul>
<li><strong>General</strong>
<ul>
<li>Small fix code snippet custom_prompt_management.md - <a href="https://github.com/BerriAI/litellm/pull/15544" target="_blank" rel="noopener noreferrer">PR #15544</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-78-5#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li>
<p><strong>Cost Tracking</strong></p>
<ul>
<li>Feat: Cost Tracking - specify a global vendor discount for costs - <a href="https://github.com/BerriAI/litellm/pull/15546" target="_blank" rel="noopener noreferrer">PR #15546</a></li>
<li>Feat: UI - Allow setting Provider Discounts on UI - <a href="https://github.com/BerriAI/litellm/pull/15550" target="_blank" rel="noopener noreferrer">PR #15550</a></li>
</ul>
</li>
<li>
<p><strong>Budgets</strong></p>
<ul>
<li>Fix: improve budget clarity - <a href="https://github.com/BerriAI/litellm/pull/15682" target="_blank" rel="noopener noreferrer">PR #15682</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements">Performance / Loadbalancing / Reliability improvements<a href="https://docs.litellm.ai/release_notes/v1-78-5#performance--loadbalancing--reliability-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements" title="Direct link to Performance / Loadbalancing / Reliability improvements">​</a></h2>
<ul>
<li>
<p><strong>Router Optimizations</strong></p>
<ul>
<li>Perf(router): use shallow copy instead of deepcopy for model aliases - 10-100x faster than deepcopy on nested dict structures - <a href="https://github.com/BerriAI/litellm/pull/15576" target="_blank" rel="noopener noreferrer">PR #15576</a></li>
<li>Perf(router): optimize string concatenation in hash generation - Improves time complexity from O(n²) to O(n) - <a href="https://github.com/BerriAI/litellm/pull/15575" target="_blank" rel="noopener noreferrer">PR #15575</a></li>
<li>Perf(router): optimize model lookups with O(1) data structures - Replace O(n) scans with index map lookups - <a href="https://github.com/BerriAI/litellm/pull/15578" target="_blank" rel="noopener noreferrer">PR #15578</a></li>
<li>Perf(router): optimize model lookups with O(1) index maps - Use model_id_to_deployment_index_map and model_name_to_deployment_indices for instant lookups - <a href="https://github.com/BerriAI/litellm/pull/15574" target="_blank" rel="noopener noreferrer">PR #15574</a></li>
<li>Perf(router): optimize timing functions in completion hot path - Use time.perf_counter() for duration measurements and time.monotonic() for timeout calculations, providing 30-40% faster timing calls - <a href="https://github.com/BerriAI/litellm/pull/15617" target="_blank" rel="noopener noreferrer">PR #15617</a></li>
</ul>
</li>
<li>
<p><strong>SSL/TLS Performance</strong></p>
<ul>
<li>Feat(ssl): add configurable ECDH curve for TLS performance - Configure via ssl_ecdh_curve setting to disable PQC on OpenSSL 3.x for better performance - <a href="https://github.com/BerriAI/litellm/pull/15617" target="_blank" rel="noopener noreferrer">PR #15617</a></li>
</ul>
</li>
<li>
<p><strong>Token Counter</strong></p>
<ul>
<li>Fix(token-counter): extract model_info from deployment for custom_tokenizer - <a href="https://github.com/BerriAI/litellm/pull/15680" target="_blank" rel="noopener noreferrer">PR #15680</a></li>
</ul>
</li>
<li>
<p><strong>Performance Metrics</strong></p>
<ul>
<li>Add: perf summary - <a href="https://github.com/BerriAI/litellm/pull/15458" target="_blank" rel="noopener noreferrer">PR #15458</a></li>
</ul>
</li>
<li>
<p><strong>CI/CD</strong></p>
<ul>
<li>Fix: CI/CD - Missing env key &amp; Linter type error - <a href="https://github.com/BerriAI/litellm/pull/15606" target="_blank" rel="noopener noreferrer">PR #15606</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates">Documentation Updates<a href="https://docs.litellm.ai/release_notes/v1-78-5#documentation-updates" class="hash-link" aria-label="Direct link to Documentation Updates" title="Direct link to Documentation Updates">​</a></h2>
<ul>
<li>
<p><strong>Provider Documentation</strong></p>
<ul>
<li>Litellm docs 10 11 2025 - <a href="https://github.com/BerriAI/litellm/pull/15457" target="_blank" rel="noopener noreferrer">PR #15457</a></li>
<li>Docs: add ecs deployment guide - <a href="https://github.com/BerriAI/litellm/pull/15468" target="_blank" rel="noopener noreferrer">PR #15468</a></li>
<li>Docs: Update benchmark results - <a href="https://github.com/BerriAI/litellm/pull/15461" target="_blank" rel="noopener noreferrer">PR #15461</a></li>
<li>Fix: add missing context to benchmark docs - <a href="https://github.com/BerriAI/litellm/pull/15688" target="_blank" rel="noopener noreferrer">PR #15688</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Fixed a few typos - <a href="https://github.com/BerriAI/litellm/pull/15267" target="_blank" rel="noopener noreferrer">PR #15267</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-78-5#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@jlan-nl made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15374" target="_blank" rel="noopener noreferrer">PR #15374</a></li>
<li>@ImadSaddik made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15267" target="_blank" rel="noopener noreferrer">PR #15267</a></li>
<li>@huangyafei made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15472" target="_blank" rel="noopener noreferrer">PR #15472</a></li>
<li>@mubashir1osmani made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15468" target="_blank" rel="noopener noreferrer">PR #15468</a></li>
<li>@kowyo made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15465" target="_blank" rel="noopener noreferrer">PR #15465</a></li>
<li>@dhruvyad made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15448" target="_blank" rel="noopener noreferrer">PR #15448</a></li>
<li>@davizucon made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15544" target="_blank" rel="noopener noreferrer">PR #15544</a></li>
<li>@FelipeRodriguesGare made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15540" target="_blank" rel="noopener noreferrer">PR #15540</a></li>
<li>@ndrsfel made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15557" target="_blank" rel="noopener noreferrer">PR #15557</a></li>
<li>@shinharaguchi made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15598" target="_blank" rel="noopener noreferrer">PR #15598</a></li>
<li>@TensorNull made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15591" target="_blank" rel="noopener noreferrer">PR #15591</a></li>
<li>@TeddyAmkie made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15583" target="_blank" rel="noopener noreferrer">PR #15583</a></li>
<li>@aniketmaurya made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15580" target="_blank" rel="noopener noreferrer">PR #15580</a></li>
<li>@eddierichter-amd made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15554" target="_blank" rel="noopener noreferrer">PR #15554</a></li>
<li>@konekohana made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15535" target="_blank" rel="noopener noreferrer">PR #15535</a></li>
<li>@Classic298 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15495" target="_blank" rel="noopener noreferrer">PR #15495</a></li>
<li>@afogel made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15599" target="_blank" rel="noopener noreferrer">PR #15599</a></li>
<li>@orolega made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15633" target="_blank" rel="noopener noreferrer">PR #15633</a></li>
<li>@LucasSugi made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15634" target="_blank" rel="noopener noreferrer">PR #15634</a></li>
<li>@uc4w6c made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15619" target="_blank" rel="noopener noreferrer">PR #15619</a></li>
<li>@Sameerlite made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15658" target="_blank" rel="noopener noreferrer">PR #15658</a></li>
<li>@yuneng-jiang made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15672" target="_blank" rel="noopener noreferrer">PR #15672</a></li>
<li>@Nikro made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15680" target="_blank" rel="noopener noreferrer">PR #15680</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog">Full Changelog<a href="https://docs.litellm.ai/release_notes/v1-78-5#full-changelog" class="hash-link" aria-label="Direct link to Full Changelog" title="Direct link to Full Changelog">​</a></h2>
<p><strong><a href="https://github.com/BerriAI/litellm/compare/v1.78.0-stable...v1.78.4-stable" target="_blank" rel="noopener noreferrer">View complete changelog on GitHub</a></strong></p>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[v1.78.0-stable - MCP Gateway: Control Tool Access by Team, Key]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-78-0</id>
        <link href="https://docs.litellm.ai/release_notes/v1-78-0"/>
        <updated>2025-10-11T10:00:00.000Z</updated>
        <summary type="html"><![CDATA[Deploy this version]]></summary>
        <content type="html"><![CDATA[<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-78-0#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">docker.litellm.ai/berriai/litellm:v1.78.0-stable</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.78.0.post1</span><br></span></code></pre></div></div></div></div></div>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-78-0#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<ul>
<li><strong>MCP Gateway - Control Tool Access by Team, Key</strong> - Control MCP tool access by team/key.</li>
<li><strong>Performance Improvements</strong> - 70% Lower p99 Latency</li>
<li><strong>GPT-5 Pro &amp; GPT-Image-1-Mini</strong> - Day 0 support for OpenAI's GPT-5 Pro (400K context) and gpt-image-1-mini image generation</li>
<li><strong>EnkryptAI Guardrails</strong> - New guardrail integration for content moderation</li>
<li><strong>Tag-Based Budgets</strong> - Support for setting budgets based on request tags</li>
</ul>
<hr>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-gateway---control-tool-access-by-team-key">MCP Gateway - Control Tool Access by Team, Key<a href="https://docs.litellm.ai/release_notes/v1-78-0#mcp-gateway---control-tool-access-by-team-key" class="hash-link" aria-label="Direct link to MCP Gateway - Control Tool Access by Team, Key" title="Direct link to MCP Gateway - Control Tool Access by Team, Key">​</a></h3>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAFCAYAAAB8ZH1oAAAACXBIWXMAACxLAAAsSwGlPZapAAAAdUlEQVR4nD2NWwoEIQwEvf81RXA+xHnoqkm0l3aZDRQkdJF28TjgvUfOGSKC3jvGGHtX1T+OQggBMcYtXVdFa5QFNic4ay042ma2YVhqQ/00iCjUDAvA/RQ4VrDqraNIntqRzrI/pnT+xBfKrOGITuS775vZF6pTwroE5VMUAAAAAElFTkSuQmCC&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="334"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/tool_control.c192a32.640.png" srcset="/assets/ideal-img/tool_control.c192a32.640.png 640w,/assets/ideal-img/tool_control.ab164aa.1920.png 1920w" width="640" height="334"></noscript></div>
<br>
<p>Proxy admins can now control MCP tool access by team or key. This makes it easy to grant different teams selective access to tools from the same MCP server.</p>
<p>For example, you can now give your Engineering team access to <code>list_repositories</code>, <code>create_issue</code>, and <code>search_code</code> tools, while Sales only gets <code>search_code</code> and <code>close_issue</code> tools.</p>
<p>This makes it easier for Proxy Admins to govern MCP Tool Access.</p>
<p><a href="https://docs.litellm.ai/docs/mcp_control#set-allowed-tools-for-a-key-team-or-organization">Get Started</a></p>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance---70-lower-p99-latency">Performance - 70% Lower p99 Latency<a href="https://docs.litellm.ai/release_notes/v1-78-0#performance---70-lower-p99-latency" class="hash-link" aria-label="Direct link to Performance - 70% Lower p99 Latency" title="Direct link to Performance - 70% Lower p99 Latency">​</a></h2>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAECAYAAAC3OK7NAAAACXBIWXMAABYlAAAWJQFJUiTwAAAAg0lEQVR4nG2MQQrCMAAE83dP/qCGYIsFfUBFvbUJpU/IWUS89dQmJW3KSOPVhYWF2V3R9z3Oe5a48E/rujLPM2KaJh7VnUtxpjyVSCnJDhnHPEcphdY6lYVzDlUX7Ko91/pG13aY1qBbQ9M0WGt/jyEE3sOHp3vhF5/WMcbkLW98HEe+nciO2qktIf0AAAAASUVORK5CYII=&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="251"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/1_78_0_perf.80a9a6d.640.png" srcset="/assets/ideal-img/1_78_0_perf.80a9a6d.640.png 640w,/assets/ideal-img/1_78_0_perf.c75037e.1920.png 1920w" width="640" height="251"></noscript></div>
<br>
<p>This release cuts p99 latency by 70% on LiteLLM AI Gateway, making it even better for low-latency use cases.</p>
<p>These gains come from two key enhancements:</p>
<p><strong>Reliable Sessions</strong></p>
<p>Added support for shared sessions with aiohttp. The shared_session parameter is now consistently used across all calls, enabling connection pooling.</p>
<p><strong>Faster Routing</strong></p>
<p>A new <code>model_name_to_deployment_indices</code> hash map replaces O(n) list scans in <code>_get_all_deployments()</code> with O(1) hash lookups, boosting routing performance and scalability.</p>
<p>As a result, performance improved across all latency percentiles:</p>
<ul>
<li><strong>Median latency:</strong> 110 ms → <strong>100 ms</strong> (−9.1%)</li>
<li><strong>p95 latency:</strong> 440 ms → <strong>150 ms</strong> (−65.9%)</li>
<li><strong>p99 latency:</strong> 810 ms → <strong>240 ms</strong> (−70.4%)</li>
<li><strong>Average latency:</strong> 310 ms → <strong>111.73 ms</strong> (−64.0%)</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="test-setup"><strong>Test Setup</strong><a href="https://docs.litellm.ai/release_notes/v1-78-0#test-setup" class="hash-link" aria-label="Direct link to test-setup" title="Direct link to test-setup">​</a></h3>
<p><strong>Locust</strong></p>
<ul>
<li><strong>Concurrent users:</strong>&nbsp;1,000</li>
<li><strong>Ramp-up:</strong>&nbsp;500</li>
</ul>
<p><strong>System Specs</strong></p>
<ul>
<li><strong>Database was used</strong></li>
<li><strong>CPU:</strong>&nbsp;4 vCPUs</li>
<li><strong>Memory:</strong>&nbsp;8 GB RAM</li>
<li><strong>LiteLLM Workers:</strong>&nbsp;4</li>
<li><strong>Instances</strong>: 4</li>
</ul>
<p><strong>Configuration (config.yaml)</strong></p>
<p>View the complete configuration:&nbsp;<a href="https://gist.github.com/AlexsanderHamir/53f7d554a5d2afcf2c4edb5b6be68ff4" target="_blank" rel="noopener noreferrer">gist.github.com/AlexsanderHamir/config.yaml</a></p>
<p><strong>Load Script (no_cache_hits.py)</strong></p>
<p>View the complete load testing script:&nbsp;<a href="https://gist.github.com/AlexsanderHamir/42c33d7a4dc7a57f56a78b560dee3a42" target="_blank" rel="noopener noreferrer">gist.github.com/AlexsanderHamir/no_cache_hits.py</a></p>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-78-0#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support">New Model Support<a href="https://docs.litellm.ai/release_notes/v1-78-0#new-model-support" class="hash-link" aria-label="Direct link to New Model Support" title="Direct link to New Model Support">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th><th>Features</th></tr></thead><tbody><tr><td>OpenAI</td><td><code>gpt-5-pro</code></td><td>400K</td><td>$15.00</td><td>$120.00</td><td>Responses API, reasoning, vision, function calling, prompt caching, web search</td></tr><tr><td>OpenAI</td><td><code>gpt-5-pro-2025-10-06</code></td><td>400K</td><td>$15.00</td><td>$120.00</td><td>Responses API, reasoning, vision, function calling, prompt caching, web search</td></tr><tr><td>OpenAI</td><td><code>gpt-image-1-mini</code></td><td>-</td><td>$2.00/img</td><td>-</td><td>Image generation and editing</td></tr><tr><td>OpenAI</td><td><code>gpt-realtime-mini</code></td><td>128K</td><td>$0.60</td><td>$2.40</td><td>Realtime audio, function calling</td></tr><tr><td>Azure AI</td><td><code>azure_ai/Phi-4-mini-reasoning</code></td><td>131K</td><td>$0.08</td><td>$0.32</td><td>Function calling</td></tr><tr><td>Azure AI</td><td><code>azure_ai/Phi-4-reasoning</code></td><td>32K</td><td>$0.125</td><td>$0.50</td><td>Function calling, reasoning</td></tr><tr><td>Azure AI</td><td><code>azure_ai/MAI-DS-R1</code></td><td>128K</td><td>$1.35</td><td>$5.40</td><td>Reasoning, function calling</td></tr><tr><td>Bedrock</td><td><code>au.anthropic.claude-sonnet-4-5-20250929-v1:0</code></td><td>200K</td><td>$3.30</td><td>$16.50</td><td>Chat, reasoning, vision, function calling, prompt caching</td></tr><tr><td>Bedrock</td><td><code>global.anthropic.claude-sonnet-4-5-20250929-v1:0</code></td><td>200K</td><td>$3.00</td><td>$15.00</td><td>Chat, reasoning, vision, function calling, prompt caching</td></tr><tr><td>Bedrock</td><td><code>global.anthropic.claude-sonnet-4-20250514-v1:0</code></td><td>1M</td><td>$3.00</td><td>$15.00</td><td>Chat, reasoning, vision, function calling, prompt caching</td></tr><tr><td>Bedrock</td><td><code>cohere.embed-v4:0</code></td><td>128K</td><td>$0.12</td><td>-</td><td>Embeddings, image input support</td></tr><tr><td>OCI</td><td><code>oci/cohere.command-latest</code></td><td>128K</td><td>$1.56</td><td>$1.56</td><td>Function calling</td></tr><tr><td>OCI</td><td><code>oci/cohere.command-a-03-2025</code></td><td>256K</td><td>$1.56</td><td>$1.56</td><td>Function calling</td></tr><tr><td>OCI</td><td><code>oci/cohere.command-plus-latest</code></td><td>128K</td><td>$1.56</td><td>$1.56</td><td>Function calling</td></tr><tr><td>Together AI</td><td><code>together_ai/moonshotai/Kimi-K2-Instruct-0905</code></td><td>262K</td><td>$1.00</td><td>$3.00</td><td>Function calling</td></tr><tr><td>Together AI</td><td><code>together_ai/Qwen/Qwen3-Next-80B-A3B-Instruct</code></td><td>262K</td><td>$0.15</td><td>$1.50</td><td>Function calling</td></tr><tr><td>Together AI</td><td><code>together_ai/Qwen/Qwen3-Next-80B-A3B-Thinking</code></td><td>262K</td><td>$0.15</td><td>$1.50</td><td>Function calling</td></tr><tr><td>Vertex AI</td><td>MedGemma models</td><td>Varies</td><td>Varies</td><td>Varies</td><td>Medical-focused Gemma models on custom endpoints</td></tr><tr><td>Watson X</td><td>27 new foundation models</td><td>Varies</td><td>Varies</td><td>Varies</td><td>Granite, Llama, Mistral families</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-78-0#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI</a></strong></p>
<ul>
<li>Add GPT-5 Pro model configuration and documentation - <a href="https://github.com/BerriAI/litellm/pull/15258" target="_blank" rel="noopener noreferrer">PR #15258</a></li>
<li>Add stop parameter to non-supported params for GPT-5 - <a href="https://github.com/BerriAI/litellm/pull/15244" target="_blank" rel="noopener noreferrer">PR #15244</a></li>
<li>Day 0 Support, Add gpt-image-1-mini - <a href="https://github.com/BerriAI/litellm/pull/15259" target="_blank" rel="noopener noreferrer">PR #15259</a></li>
<li>Add gpt-realtime-mini support - <a href="https://github.com/BerriAI/litellm/pull/15283" target="_blank" rel="noopener noreferrer">PR #15283</a></li>
<li>Add gpt-5-pro-2025-10-06 to model costs - <a href="https://github.com/BerriAI/litellm/pull/15344" target="_blank" rel="noopener noreferrer">PR #15344</a></li>
<li>Minimal fix: gpt5 models should not go on cooldown when called with temperature!=1 - <a href="https://github.com/BerriAI/litellm/pull/15330" target="_blank" rel="noopener noreferrer">PR #15330</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/snowflake">Snowflake Cortex</a></strong></p>
<ul>
<li>Add function calling support for Snowflake Cortex REST API - <a href="https://github.com/BerriAI/litellm/pull/15221" target="_blank" rel="noopener noreferrer">PR #15221</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Gemini</a></strong></p>
<ul>
<li>Fix header forwarding for Gemini/Vertex AI providers in proxy mode - <a href="https://github.com/BerriAI/litellm/pull/15231" target="_blank" rel="noopener noreferrer">PR #15231</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/azure">Azure</a></strong></p>
<ul>
<li>Removed stop param from unsupported azure models - <a href="https://github.com/BerriAI/litellm/pull/15229" target="_blank" rel="noopener noreferrer">PR #15229</a></li>
<li>Fix(azure/responses): remove invalid status param from azure call - <a href="https://github.com/BerriAI/litellm/pull/15253" target="_blank" rel="noopener noreferrer">PR #15253</a></li>
<li>Add new Azure AI models with pricing details - <a href="https://github.com/BerriAI/litellm/pull/15387" target="_blank" rel="noopener noreferrer">PR #15387</a></li>
<li>AzureAD Default credentials - select credential type based on environment - <a href="https://github.com/BerriAI/litellm/pull/14470" target="_blank" rel="noopener noreferrer">PR #14470</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong></p>
<ul>
<li>Add Global Cross-Region Inference - <a href="https://github.com/BerriAI/litellm/pull/15210" target="_blank" rel="noopener noreferrer">PR #15210</a></li>
<li>Add Cohere Embed v4 support for AWS Bedrock - <a href="https://github.com/BerriAI/litellm/pull/15298" target="_blank" rel="noopener noreferrer">PR #15298</a></li>
<li>Fix(bedrock): include cacheWriteInputTokens in prompt_tokens calculation - <a href="https://github.com/BerriAI/litellm/pull/15292" target="_blank" rel="noopener noreferrer">PR #15292</a></li>
<li>Add Bedrock AU Cross-Region Inference for Claude Sonnet 4.5 - <a href="https://github.com/BerriAI/litellm/pull/15402" target="_blank" rel="noopener noreferrer">PR #15402</a></li>
<li>Converse → /v1/messages streaming doesn't handle parallel tool calls with Claude models - <a href="https://github.com/BerriAI/litellm/pull/15315" target="_blank" rel="noopener noreferrer">PR #15315</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/vertex">Vertex AI</a></strong></p>
<ul>
<li>Implement Context Caching for Vertex AI provider - <a href="https://github.com/BerriAI/litellm/pull/15226" target="_blank" rel="noopener noreferrer">PR #15226</a></li>
<li>Support for Vertex AI Gemma Models on Custom Endpoints - <a href="https://github.com/BerriAI/litellm/pull/15397" target="_blank" rel="noopener noreferrer">PR #15397</a></li>
<li>VertexAI - gemma model family support (custom endpoints) - <a href="https://github.com/BerriAI/litellm/pull/15419" target="_blank" rel="noopener noreferrer">PR #15419</a></li>
<li>VertexAI Gemma model family streaming support + Added MedGemma - <a href="https://github.com/BerriAI/litellm/pull/15427" target="_blank" rel="noopener noreferrer">PR #15427</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/oci">OCI</a></strong></p>
<ul>
<li>Add OCI Cohere support with tool calling and streaming capabilities - <a href="https://github.com/BerriAI/litellm/pull/15365" target="_blank" rel="noopener noreferrer">PR #15365</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/watsonx">Watson X</a></strong></p>
<ul>
<li>Add Watson X foundation model definitions to model_prices_and_context_window.json - <a href="https://github.com/BerriAI/litellm/pull/15219" target="_blank" rel="noopener noreferrer">PR #15219</a></li>
<li>Watsonx - Apply correct prompt templates for openai/gpt-oss model family - <a href="https://github.com/BerriAI/litellm/pull/15341" target="_blank" rel="noopener noreferrer">PR #15341</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/openrouter">OpenRouter</a></strong></p>
<ul>
<li>Fix - (openrouter): move cache_control to content blocks for claude/gemini - <a href="https://github.com/BerriAI/litellm/pull/15345" target="_blank" rel="noopener noreferrer">PR #15345</a></li>
<li>Fix - OpenRouter cache_control to only apply to last content block - <a href="https://github.com/BerriAI/litellm/pull/15395" target="_blank" rel="noopener noreferrer">PR #15395</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/togetherai">Together AI</a></strong></p>
<ul>
<li>Add new together models - <a href="https://github.com/BerriAI/litellm/pull/15383" target="_blank" rel="noopener noreferrer">PR #15383</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-78-0#bug-fixes" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li><strong>General</strong>
<ul>
<li>Bug fix: gpt-5-chat-latest has incorrect max_input_tokens value - <a href="https://github.com/BerriAI/litellm/pull/15116" target="_blank" rel="noopener noreferrer">PR #15116</a></li>
<li>Fix reasoning response ID - <a href="https://github.com/BerriAI/litellm/pull/15265" target="_blank" rel="noopener noreferrer">PR #15265</a></li>
<li>Fix issue with parsing assistant messages - <a href="https://github.com/BerriAI/litellm/pull/15320" target="_blank" rel="noopener noreferrer">PR #15320</a></li>
<li>Fix litellm_param based costing - <a href="https://github.com/BerriAI/litellm/pull/15336" target="_blank" rel="noopener noreferrer">PR #15336</a></li>
<li>Fix lint errors - <a href="https://github.com/BerriAI/litellm/pull/15406" target="_blank" rel="noopener noreferrer">PR #15406</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-78-0#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-78-0#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/response_api">Responses API</a></strong></p>
<ul>
<li>Added streaming support for response api streaming image generation - <a href="https://github.com/BerriAI/litellm/pull/15269" target="_blank" rel="noopener noreferrer">PR #15269</a></li>
<li>Add native Responses API support for litellm_proxy provider - <a href="https://github.com/BerriAI/litellm/pull/15347" target="_blank" rel="noopener noreferrer">PR #15347</a></li>
<li>Temporarily relax ResponsesAPIResponse parsing to support custom backends (e.g., vLLM) - <a href="https://github.com/BerriAI/litellm/pull/15362" target="_blank" rel="noopener noreferrer">PR #15362</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/files_api">Files API</a></strong></p>
<ul>
<li>Feat(files): add @client decorator to file operations - <a href="https://github.com/BerriAI/litellm/pull/15339" target="_blank" rel="noopener noreferrer">PR #15339</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">/generateContent</a></strong></p>
<ul>
<li>Fix gemini cli by actually streaming the response - <a href="https://github.com/BerriAI/litellm/pull/15264" target="_blank" rel="noopener noreferrer">PR #15264</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/pass_through/azure">Azure Passthrough</a></strong></p>
<ul>
<li>Azure - passthrough support with router models - <a href="https://github.com/BerriAI/litellm/pull/15240" target="_blank" rel="noopener noreferrer">PR #15240</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs">Bugs<a href="https://docs.litellm.ai/release_notes/v1-78-0#bugs" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>General</strong>
<ul>
<li>Fix x-litellm-cache-key header not being returned on cache hit - <a href="https://github.com/BerriAI/litellm/pull/15348" target="_blank" rel="noopener noreferrer">PR #15348</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-78-0#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-78-0#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong>Proxy CLI Auth</strong></p>
<ul>
<li>Proxy CLI - dont store existing key in the URL, store it in the state param - <a href="https://github.com/BerriAI/litellm/pull/15290" target="_blank" rel="noopener noreferrer">PR #15290</a></li>
</ul>
</li>
<li>
<p><strong>Models + Endpoints</strong></p>
<ul>
<li>Make PATCH <code>/model/{model_id}/update</code> handle <code>team_id</code> consistently with POST <code>/model/new</code> - <a href="https://github.com/BerriAI/litellm/pull/15297" target="_blank" rel="noopener noreferrer">PR #15297</a></li>
<li>Feature: adds Infinity as a provider in the UI - <a href="https://github.com/BerriAI/litellm/pull/15285" target="_blank" rel="noopener noreferrer">PR #15285</a></li>
<li>Fix: model + endpoints page crash when config file contains router_settings.model_group_alias - <a href="https://github.com/BerriAI/litellm/pull/15308" target="_blank" rel="noopener noreferrer">PR #15308</a></li>
<li>Models &amp; Endpoints Initial Refactor - <a href="https://github.com/BerriAI/litellm/pull/15435" target="_blank" rel="noopener noreferrer">PR #15435</a></li>
<li>Litellm UI API Reference page updates - <a href="https://github.com/BerriAI/litellm/pull/15438" target="_blank" rel="noopener noreferrer">PR #15438</a></li>
</ul>
</li>
<li>
<p><strong>Teams</strong></p>
<ul>
<li>Teams page: new column "Your Role" on the teams table - <a href="https://github.com/BerriAI/litellm/pull/15384" target="_blank" rel="noopener noreferrer">PR #15384</a></li>
<li>LiteLLM Dashboard Teams UI refactor - <a href="https://github.com/BerriAI/litellm/pull/15418" target="_blank" rel="noopener noreferrer">PR #15418</a></li>
</ul>
</li>
<li>
<p><strong>UI Infrastructure</strong></p>
<ul>
<li>Added prettier to autoformat frontend - <a href="https://github.com/BerriAI/litellm/pull/15215" target="_blank" rel="noopener noreferrer">PR #15215</a></li>
<li>Adds turbopack to the npm run dev command in UI to build faster during development - <a href="https://github.com/BerriAI/litellm/pull/15250" target="_blank" rel="noopener noreferrer">PR #15250</a></li>
<li>(perf) fix: Replaces bloated key list calls with lean key aliases endpoint - <a href="https://github.com/BerriAI/litellm/pull/15252" target="_blank" rel="noopener noreferrer">PR #15252</a></li>
<li>Potentially fixes a UI spasm issue with an expired cookie - <a href="https://github.com/BerriAI/litellm/pull/15309" target="_blank" rel="noopener noreferrer">PR #15309</a></li>
<li>LiteLLM UI Refactor Infrastructure - <a href="https://github.com/BerriAI/litellm/pull/15236" target="_blank" rel="noopener noreferrer">PR #15236</a></li>
<li>Enforces removal of unused imports from UI - <a href="https://github.com/BerriAI/litellm/pull/15416" target="_blank" rel="noopener noreferrer">PR #15416</a></li>
<li>Fix: usage page &gt;&gt; Model Activity &gt;&gt; spend per day graph: y-axis clipping on large spend values - <a href="https://github.com/BerriAI/litellm/pull/15389" target="_blank" rel="noopener noreferrer">PR #15389</a></li>
<li>Updates guardrail provider logos - <a href="https://github.com/BerriAI/litellm/pull/15421" target="_blank" rel="noopener noreferrer">PR #15421</a></li>
</ul>
</li>
<li>
<p><strong>Admin Settings</strong></p>
<ul>
<li>Fix: Router settings do not update despite success message - <a href="https://github.com/BerriAI/litellm/pull/15249" target="_blank" rel="noopener noreferrer">PR #15249</a></li>
<li>Fix: Prevents DB from accidentally overriding config file values if they are empty in DB - <a href="https://github.com/BerriAI/litellm/pull/15340" target="_blank" rel="noopener noreferrer">PR #15340</a></li>
</ul>
</li>
<li>
<p><strong>SSO</strong></p>
<ul>
<li>SSO - support EntraID app roles - <a href="https://github.com/BerriAI/litellm/pull/15351" target="_blank" rel="noopener noreferrer">PR #15351</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="logging--guardrail--prompt-management-integrations">Logging / Guardrail / Prompt Management Integrations<a href="https://docs.litellm.ai/release_notes/v1-78-0#logging--guardrail--prompt-management-integrations" class="hash-link" aria-label="Direct link to Logging / Guardrail / Prompt Management Integrations" title="Direct link to Logging / Guardrail / Prompt Management Integrations">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-3">Features<a href="https://docs.litellm.ai/release_notes/v1-78-0#features-3" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/observability/posthog">PostHog</a></strong>
<ul>
<li>Feat: posthog per request api key - <a href="https://github.com/BerriAI/litellm/pull/15379" target="_blank" rel="noopener noreferrer">PR #15379</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails">Guardrails<a href="https://docs.litellm.ai/release_notes/v1-78-0#guardrails" class="hash-link" aria-label="Direct link to Guardrails" title="Direct link to Guardrails">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">EnkryptAI</a></strong>
<ul>
<li>Add EnkryptAI Guardrails on LiteLLM - <a href="https://github.com/BerriAI/litellm/pull/15390" target="_blank" rel="noopener noreferrer">PR #15390</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-78-0#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li>
<p><strong>Tag Management</strong></p>
<ul>
<li>Tag Management - Add support for setting tag based budgets - <a href="https://github.com/BerriAI/litellm/pull/15433" target="_blank" rel="noopener noreferrer">PR #15433</a></li>
</ul>
</li>
<li>
<p><strong>Dynamic Rate Limiter v3</strong></p>
<ul>
<li>QA/Fixes - Dynamic Rate Limiter v3 - final QA - <a href="https://github.com/BerriAI/litellm/pull/15311" target="_blank" rel="noopener noreferrer">PR #15311</a></li>
<li>Fix dynamic Rate limiter v3 - inserting litellm_model_saturation - <a href="https://github.com/BerriAI/litellm/pull/15394" target="_blank" rel="noopener noreferrer">PR #15394</a></li>
</ul>
</li>
<li>
<p><strong>Shared Health Check</strong></p>
<ul>
<li>Implement Shared Health Check State Across Pods - <a href="https://github.com/BerriAI/litellm/pull/15380" target="_blank" rel="noopener noreferrer">PR #15380</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-gateway">MCP Gateway<a href="https://docs.litellm.ai/release_notes/v1-78-0#mcp-gateway" class="hash-link" aria-label="Direct link to MCP Gateway" title="Direct link to MCP Gateway">​</a></h2>
<ul>
<li>
<p><strong>Tool Control</strong></p>
<ul>
<li>MCP Gateway - UI - Select allowed tools for Key, Teams - <a href="https://github.com/BerriAI/litellm/pull/15241" target="_blank" rel="noopener noreferrer">PR #15241</a></li>
<li>MCP Gateway - Backend - Allow storing allowed tools by team/key - <a href="https://github.com/BerriAI/litellm/pull/15243" target="_blank" rel="noopener noreferrer">PR #15243</a></li>
<li>MCP Gateway - Fine-grained Database Object Storage Control - <a href="https://github.com/BerriAI/litellm/pull/15255" target="_blank" rel="noopener noreferrer">PR #15255</a></li>
<li>MCP Gateway - Litellm mcp fixes team control - <a href="https://github.com/BerriAI/litellm/pull/15304" target="_blank" rel="noopener noreferrer">PR #15304</a></li>
<li>MCP Gateway - QA/Fixes - Ensure Team/Key level enforcement works for MCPs - <a href="https://github.com/BerriAI/litellm/pull/15305" target="_blank" rel="noopener noreferrer">PR #15305</a></li>
<li>Feature: Include server_name in /v1/mcp/server/health endpoint response - <a href="https://github.com/BerriAI/litellm/pull/15431" target="_blank" rel="noopener noreferrer">PR #15431</a></li>
</ul>
</li>
<li>
<p><strong>OpenAPI Integration</strong></p>
<ul>
<li>MCP - support converting OpenAPI specs to MCP servers - <a href="https://github.com/BerriAI/litellm/pull/15343" target="_blank" rel="noopener noreferrer">PR #15343</a></li>
<li>MCP - specify allowed params per tool - <a href="https://github.com/BerriAI/litellm/pull/15346" target="_blank" rel="noopener noreferrer">PR #15346</a></li>
</ul>
</li>
<li>
<p><strong>Configuration</strong></p>
<ul>
<li>MCP - support setting CA_BUNDLE_PATH - <a href="https://github.com/BerriAI/litellm/pull/15253" target="_blank" rel="noopener noreferrer">PR #15253</a></li>
<li>Fix: Ensure MCP client stays open during tool call - <a href="https://github.com/BerriAI/litellm/pull/15391" target="_blank" rel="noopener noreferrer">PR #15391</a></li>
<li>Remove hardcoded "public" schema in migration.sql - <a href="https://github.com/BerriAI/litellm/pull/15363" target="_blank" rel="noopener noreferrer">PR #15363</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements">Performance / Loadbalancing / Reliability improvements<a href="https://docs.litellm.ai/release_notes/v1-78-0#performance--loadbalancing--reliability-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements" title="Direct link to Performance / Loadbalancing / Reliability improvements">​</a></h2>
<ul>
<li>
<p><strong>Router Optimizations</strong></p>
<ul>
<li>Fix - Router: add model_name index for O(1) deployment lookups - <a href="https://github.com/BerriAI/litellm/pull/15113" target="_blank" rel="noopener noreferrer">PR #15113</a></li>
<li>Refactor Utils: extract inner function from client - <a href="https://github.com/BerriAI/litellm/pull/15234" target="_blank" rel="noopener noreferrer">PR #15234</a></li>
<li>Fix Networking: remove limitations - <a href="https://github.com/BerriAI/litellm/pull/15302" target="_blank" rel="noopener noreferrer">PR #15302</a></li>
</ul>
</li>
<li>
<p><strong>Session Management</strong></p>
<ul>
<li>Fix - Sessions not being shared - <a href="https://github.com/BerriAI/litellm/pull/15388" target="_blank" rel="noopener noreferrer">PR #15388</a></li>
<li>Fix: remove panic from hot path - <a href="https://github.com/BerriAI/litellm/pull/15396" target="_blank" rel="noopener noreferrer">PR #15396</a></li>
<li>Fix - shared session parsing and usage issue - <a href="https://github.com/BerriAI/litellm/pull/15440" target="_blank" rel="noopener noreferrer">PR #15440</a></li>
<li>Fix: handle closed aiohttp sessions - <a href="https://github.com/BerriAI/litellm/pull/15442" target="_blank" rel="noopener noreferrer">PR #15442</a></li>
<li>Fix: prevent session leaks when recreating aiohttp sessions - <a href="https://github.com/BerriAI/litellm/pull/15443" target="_blank" rel="noopener noreferrer">PR #15443</a></li>
</ul>
</li>
<li>
<p><strong>SSL/TLS Performance</strong></p>
<ul>
<li>Perf: optimize SSL/TLS handshake performance with prioritized cipher - <a href="https://github.com/BerriAI/litellm/pull/15398" target="_blank" rel="noopener noreferrer">PR #15398</a></li>
</ul>
</li>
<li>
<p><strong>Dependencies</strong></p>
<ul>
<li>Upgrades tenacity version to 8.5.0 - <a href="https://github.com/BerriAI/litellm/pull/15303" target="_blank" rel="noopener noreferrer">PR #15303</a></li>
</ul>
</li>
<li>
<p><strong>Data Masking</strong></p>
<ul>
<li>Fix - SensitiveDataMasker converts lists to string - <a href="https://github.com/BerriAI/litellm/pull/15420" target="_blank" rel="noopener noreferrer">PR #15420</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="general-ai-gateway-improvements">General AI Gateway Improvements<a href="https://docs.litellm.ai/release_notes/v1-78-0#general-ai-gateway-improvements" class="hash-link" aria-label="Direct link to General AI Gateway Improvements" title="Direct link to General AI Gateway Improvements">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="security">Security<a href="https://docs.litellm.ai/release_notes/v1-78-0#security" class="hash-link" aria-label="Direct link to Security" title="Direct link to Security">​</a></h4>
<ul>
<li><strong>General</strong>
<ul>
<li>Fix: redact AWS credentials when redact_user_api_key_info enabled - <a href="https://github.com/BerriAI/litellm/pull/15321" target="_blank" rel="noopener noreferrer">PR #15321</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates">Documentation Updates<a href="https://docs.litellm.ai/release_notes/v1-78-0#documentation-updates" class="hash-link" aria-label="Direct link to Documentation Updates" title="Direct link to Documentation Updates">​</a></h2>
<ul>
<li>
<p><strong>Provider Documentation</strong></p>
<ul>
<li>Update doc: perf update - <a href="https://github.com/BerriAI/litellm/pull/15211" target="_blank" rel="noopener noreferrer">PR #15211</a></li>
<li>Add W&amp;B Inference documentation - <a href="https://github.com/BerriAI/litellm/pull/15278" target="_blank" rel="noopener noreferrer">PR #15278</a></li>
</ul>
</li>
<li>
<p><strong>Deployment</strong></p>
<ul>
<li>Deletion of docker-compose buggy comment that cause <code>config.yaml</code> based startup fail - <a href="https://github.com/BerriAI/litellm/pull/15425" target="_blank" rel="noopener noreferrer">PR #15425</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-78-0#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@Gal-bloch made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15219" target="_blank" rel="noopener noreferrer">PR #15219</a></li>
<li>@lcfyi made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15315" target="_blank" rel="noopener noreferrer">PR #15315</a></li>
<li>@ashengstd made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15362" target="_blank" rel="noopener noreferrer">PR #15362</a></li>
<li>@vkolehmainen made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15363" target="_blank" rel="noopener noreferrer">PR #15363</a></li>
<li>@jlan-nl made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15330" target="_blank" rel="noopener noreferrer">PR #15330</a></li>
<li>@BCook98 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15402" target="_blank" rel="noopener noreferrer">PR #15402</a></li>
<li>@PabloGmz96 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15425" target="_blank" rel="noopener noreferrer">PR #15425</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog"><strong><a href="https://github.com/BerriAI/litellm/compare/v1.77.7.rc.1...v1.78.0.rc.1" target="_blank" rel="noopener noreferrer">Full Changelog</a></strong><a href="https://docs.litellm.ai/release_notes/v1-78-0#full-changelog" class="hash-link" aria-label="Direct link to full-changelog" title="Direct link to full-changelog">​</a></h2>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[v1.77.7-stable - 2.9x Lower Median Latency]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-77-7</id>
        <link href="https://docs.litellm.ai/release_notes/v1-77-7"/>
        <updated>2025-10-04T10:00:00.000Z</updated>
        <summary type="html"><![CDATA[Deploy this version]]></summary>
        <content type="html"><![CDATA[<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-77-7#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">docker.litellm.ai/berriai/litellm:v1.77.7.rc.1</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.77.7.rc.1</span><br></span></code></pre></div></div></div></div></div>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-77-7#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<ul>
<li><strong>Dynamic Rate Limiter v3</strong> - Automatically maximizes throughput when capacity is available (&lt; 80% saturation) by allowing lower-priority requests to use unused capacity, then switches to fair priority-based allocation under high load (≥ 80%) to prevent blocking</li>
<li><strong>Major Performance Improvements</strong> - 2.9x lower median latency at 1,000 concurrent users.</li>
<li><strong>Claude Sonnet 4.5</strong> - Support for Anthropic's new Claude Sonnet 4.5 model family with 200K+ context and tiered pricing</li>
<li><strong>MCP Gateway Enhancements</strong> - Fine-grained tool control, server permissions, and forwardable headers</li>
<li><strong>AMD Lemonade &amp; Nvidia NIM</strong> - New provider support for AMD Lemonade and Nvidia NIM Rerank</li>
<li><strong>GitLab Prompt Management</strong> - GitLab-based prompt management integration</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="performance---29x-lower-median-latency">Performance - 2.9x Lower Median Latency<a href="https://docs.litellm.ai/release_notes/v1-77-7#performance---29x-lower-median-latency" class="hash-link" aria-label="Direct link to Performance - 2.9x Lower Median Latency" title="Direct link to Performance - 2.9x Lower Median Latency">​</a></h3>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAICAYAAADA+m62AAAACXBIWXMAACxLAAAsSwGlPZapAAAA1klEQVR4nEWQzY6CMBDHea9NNFyU1+B5Nln36EN40n2A5SKuJzxzwWCLhO+Q0ALCfzNjqkl/7UznPx+tpbWGEAJJkjBxHDNkSylR1zXGcYRFwjzPURQFoihCGIZMWZaoqgpt2z6FtA3D8Mokm+j7njF3Vtd17Hx/bbD8WMBxHKzWK9i2Ddd1YeIWZWmlcbr8Ye/9wD/7+D17OJ58BEHAVbmi6hV0o7C7HvCZbPGYH8BMa8Y0Te/WZp4szyDvEmmaQtwEnwQ9iOIsNNAPKKXQNA3PRr5p/Q+kxCQByQJIdgAAAABJRU5ErkJggg==&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="488"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/perf_77_7.ea967fe.640.png" srcset="/assets/ideal-img/perf_77_7.ea967fe.640.png 640w,/assets/ideal-img/perf_77_7.4744f3d.1920.png 1920w" width="640" height="488"></noscript></div>
<br>
<p>This update removes LiteLLM router inefficiencies, reducing complexity from O(M×N) to O(1). Previously, it built a new array and ran repeated checks like data["model"] in llm_router.get_model_ids(). Now, a direct ID-to-deployment map eliminates redundant allocations and scans.</p>
<p>As a result, performance improved across all latency percentiles:</p>
<ul>
<li><strong>Median latency:</strong> 320 ms → <strong>110 ms</strong> (−65.6%)</li>
<li><strong>p95 latency:</strong> 850 ms → <strong>440 ms</strong> (−48.2%)</li>
<li><strong>p99 latency:</strong> 1,400 ms → <strong>810 ms</strong> (−42.1%)</li>
<li><strong>Average latency:</strong> 864 ms → <strong>310 ms</strong> (−64%)</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="test-setup">Test Setup<a href="https://docs.litellm.ai/release_notes/v1-77-7#test-setup" class="hash-link" aria-label="Direct link to Test Setup" title="Direct link to Test Setup">​</a></h4>
<p><strong>Locust</strong></p>
<ul>
<li><strong>Concurrent users:</strong> 1,000</li>
<li><strong>Ramp-up:</strong> 500</li>
</ul>
<p><strong>System Specs</strong></p>
<ul>
<li><strong>CPU:</strong> 4 vCPUs</li>
<li><strong>Memory:</strong> 8 GB RAM</li>
<li><strong>LiteLLM Workers:</strong> 4</li>
<li><strong>Instances</strong>: 4</li>
</ul>
<p><strong>Configuration (config.yaml)</strong></p>
<p>View the complete configuration: <a href="https://gist.github.com/AlexsanderHamir/53f7d554a5d2afcf2c4edb5b6be68ff4" target="_blank" rel="noopener noreferrer">gist.github.com/AlexsanderHamir/config.yaml</a></p>
<p><strong>Load Script (no_cache_hits.py)</strong></p>
<p>View the complete load testing script: <a href="https://gist.github.com/AlexsanderHamir/42c33d7a4dc7a57f56a78b560dee3a42" target="_blank" rel="noopener noreferrer">gist.github.com/AlexsanderHamir/no_cache_hits.py</a></p>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-oauth-20-support">MCP OAuth 2.0 Support<a href="https://docs.litellm.ai/release_notes/v1-77-7#mcp-oauth-20-support" class="hash-link" aria-label="Direct link to MCP OAuth 2.0 Support" title="Direct link to MCP OAuth 2.0 Support">​</a></h3>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/jpeg;base64,/9j/2wBDAAYEBQYFBAYGBQYHBwYIChAKCgkJChQODwwQFxQYGBcUFhYaHSUfGhsjHBYWICwgIyYnKSopGR8tMC0oMCUoKSj/2wBDAQcHBwoIChMKChMoGhYaKCgoKCgoKCgoKCgoKCgoKCgoKCgoKCgoKCgoKCgoKCgoKCgoKCgoKCgoKCgoKCgoKCj/wAARCAAFAAoDASIAAhEBAxEB/8QAFgABAQEAAAAAAAAAAAAAAAAAAAQI/8QAHhAAAQQCAwEAAAAAAAAAAAAAAQACAwQRMSEjQWH/xAAVAQEBAAAAAAAAAAAAAAAAAAACBf/EABQRAQAAAAAAAAAAAAAAAAAAAAD/2gAMAwEAAhEDEQA/AM4i50ztlr15ZZCXGZ7SXgkec4+62oiMEhEVUX//2Q==&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="334"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/mcp_updates.33d55db.640.jpg" srcset="/assets/ideal-img/mcp_updates.33d55db.640.jpg 640w,/assets/ideal-img/mcp_updates.b70bfe0.1920.jpg 1920w" width="640" height="334"></noscript></div>
<br>
<p>This release adds support for OAuth 2.0 Client Credentials for MCP servers. This is great for <strong>Internal Dev Tools</strong> use-cases, as it enables your users to call MCP servers, with their own credentials. E.g. Allowing your developers to call the Github MCP, with their own credentials.</p>
<p><a href="https://docs.litellm.ai/docs/tutorials/claude_responses_api#connecting-mcp-servers">Set it up today on Claude Code</a></p>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="scheduled-key-rotations">Scheduled Key Rotations<a href="https://docs.litellm.ai/release_notes/v1-77-7#scheduled-key-rotations" class="hash-link" aria-label="Direct link to Scheduled Key Rotations" title="Direct link to Scheduled Key Rotations">​</a></h3>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAFCAYAAAB8ZH1oAAAACXBIWXMAACxLAAAsSwGlPZapAAAAeUlEQVR4nE3NywoDIQwFUP//E2cxW1Eq46B2zENuuS7aBkIgOUlCSgk5Z8w5ISLoY0BUoKqYUxYA9o9QSkGMEbXWjft4CGDmULUNzfwMZgZ3Byuv9M7LyjljQ3c/A98xidQMfbzxui+0Z3D5Bzf4S1FFa0SOtdYXfgC32cKzrWPOlQAAAABJRU5ErkJggg==&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="331"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/schedule_key_rotations.2680e85.640.png" srcset="/assets/ideal-img/schedule_key_rotations.2680e85.640.png 640w,/assets/ideal-img/schedule_key_rotations.7b2dfb8.1920.png 1920w" width="640" height="331"></noscript></div>
<br>
<p>This release brings support for scheduling virtual key rotations on LiteLLM AI Gateway.</p>
<p>From this release you can enforce Virtual Keys to rotate on a schedule of your choice e.g every 15 days/30 days/60 days etc.</p>
<p>This is great for Proxy Admins who need to enforce security policies for production workloads.</p>
<p><a href="https://docs.litellm.ai/docs/proxy/virtual_keys#scheduled-key-rotations">Get Started</a></p>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-77-7#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support">New Model Support<a href="https://docs.litellm.ai/release_notes/v1-77-7#new-model-support" class="hash-link" aria-label="Direct link to New Model Support" title="Direct link to New Model Support">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th><th>Features</th></tr></thead><tbody><tr><td>Anthropic</td><td><code>claude-sonnet-4-5</code></td><td>200K</td><td>$3.00</td><td>$15.00</td><td>Chat, reasoning, vision, function calling, prompt caching</td></tr><tr><td>Anthropic</td><td><code>claude-sonnet-4-5-20250929</code></td><td>200K</td><td>$3.00</td><td>$15.00</td><td>Chat, reasoning, vision, function calling, prompt caching</td></tr><tr><td>Bedrock</td><td><code>eu.anthropic.claude-sonnet-4-5-20250929-v1:0</code></td><td>200K</td><td>$3.00</td><td>$15.00</td><td>Chat, reasoning, vision, function calling, prompt caching</td></tr><tr><td>Azure AI</td><td><code>azure_ai/grok-4</code></td><td>131K</td><td>$5.50</td><td>$27.50</td><td>Chat, reasoning, function calling, web search</td></tr><tr><td>Azure AI</td><td><code>azure_ai/grok-4-fast-reasoning</code></td><td>131K</td><td>$0.43</td><td>$1.73</td><td>Chat, reasoning, function calling, web search</td></tr><tr><td>Azure AI</td><td><code>azure_ai/grok-4-fast-non-reasoning</code></td><td>131K</td><td>$0.43</td><td>$1.73</td><td>Chat, function calling, web search</td></tr><tr><td>Azure AI</td><td><code>azure_ai/grok-code-fast-1</code></td><td>131K</td><td>$3.50</td><td>$17.50</td><td>Chat, function calling, web search</td></tr><tr><td>Groq</td><td><code>groq/moonshotai/kimi-k2-instruct-0905</code></td><td>Context varies</td><td>Pricing varies</td><td>Pricing varies</td><td>Chat, function calling</td></tr><tr><td>Ollama</td><td>Ollama Cloud models</td><td>Varies</td><td>Free</td><td>Free</td><td>Self-hosted models via Ollama Cloud</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-77-7#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong>
<ul>
<li>Add new claude-sonnet-4-5 model family with tiered pricing above 200K tokens - <a href="https://github.com/BerriAI/litellm/pull/15041" target="_blank" rel="noopener noreferrer">PR #15041</a></li>
<li>Add anthropic/claude-sonnet-4-5 to model price json with prompt caching support - <a href="https://github.com/BerriAI/litellm/pull/15049" target="_blank" rel="noopener noreferrer">PR #15049</a></li>
<li>Add 200K prices for Sonnet 4.5 - <a href="https://github.com/BerriAI/litellm/pull/15140" target="_blank" rel="noopener noreferrer">PR #15140</a></li>
<li>Add cost tracking for /v1/messages in streaming response - <a href="https://github.com/BerriAI/litellm/pull/15102" target="_blank" rel="noopener noreferrer">PR #15102</a></li>
<li>Add /v1/messages/count_tokens to Anthropic routes for non-admin user access - <a href="https://github.com/BerriAI/litellm/pull/15034" target="_blank" rel="noopener noreferrer">PR #15034</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Gemini</a></strong>
<ul>
<li>Ignore type param for gemini tools - <a href="https://github.com/BerriAI/litellm/pull/15022" target="_blank" rel="noopener noreferrer">PR #15022</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/vertex">Vertex AI</a></strong>
<ul>
<li>Add LiteLLM Overhead metric for VertexAI - <a href="https://github.com/BerriAI/litellm/pull/15040" target="_blank" rel="noopener noreferrer">PR #15040</a></li>
<li>Support googlemap grounding in vertex ai - <a href="https://github.com/BerriAI/litellm/pull/15179" target="_blank" rel="noopener noreferrer">PR #15179</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/azure">Azure</a></strong>
<ul>
<li>Add azure_ai grok-4 model family - <a href="https://github.com/BerriAI/litellm/pull/15137" target="_blank" rel="noopener noreferrer">PR #15137</a></li>
<li>Use the <code>extra_query</code> parameter for GET requests in Azure Batch - <a href="https://github.com/BerriAI/litellm/pull/14997" target="_blank" rel="noopener noreferrer">PR #14997</a></li>
<li>Use extra_query for download results (Batch API) - <a href="https://github.com/BerriAI/litellm/pull/15025" target="_blank" rel="noopener noreferrer">PR #15025</a></li>
<li>Add support for Azure AD token-based authorization - <a href="https://github.com/BerriAI/litellm/pull/14813" target="_blank" rel="noopener noreferrer">PR #14813</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/ollama">Ollama</a></strong>
<ul>
<li>Add ollama cloud models - <a href="https://github.com/BerriAI/litellm/pull/15008" target="_blank" rel="noopener noreferrer">PR #15008</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/groq">Groq</a></strong>
<ul>
<li>Add groq/moonshotai/kimi-k2-instruct-0905 - <a href="https://github.com/BerriAI/litellm/pull/15079" target="_blank" rel="noopener noreferrer">PR #15079</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI</a></strong>
<ul>
<li>Add support for GPT 5 codex models - <a href="https://github.com/BerriAI/litellm/pull/14841" target="_blank" rel="noopener noreferrer">PR #14841</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/deepinfra">DeepInfra</a></strong>
<ul>
<li>Update DeepInfra model data refresh with latest pricing - <a href="https://github.com/BerriAI/litellm/pull/14939" target="_blank" rel="noopener noreferrer">PR #14939</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong>
<ul>
<li>Add JP Cross-Region Inference - <a href="https://github.com/BerriAI/litellm/pull/15188" target="_blank" rel="noopener noreferrer">PR #15188</a></li>
<li>Add "eu.anthropic.claude-sonnet-4-5-20250929-v1:0" - <a href="https://github.com/BerriAI/litellm/pull/15181" target="_blank" rel="noopener noreferrer">PR #15181</a></li>
<li>Add twelvelabs bedrock Async Invoke Support - <a href="https://github.com/BerriAI/litellm/pull/14871" target="_blank" rel="noopener noreferrer">PR #14871</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/nvidia_nim">Nvidia NIM</a></strong>
<ul>
<li>Add Nvidia NIM Rerank Support - <a href="https://github.com/BerriAI/litellm/pull/15152" target="_blank" rel="noopener noreferrer">PR #15152</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-77-7#bug-fixes" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/providers/vllm">VLLM</a></strong>
<ul>
<li>Fix response_format bug in hosted vllm audio_transcription - <a href="https://github.com/BerriAI/litellm/pull/15010" target="_blank" rel="noopener noreferrer">PR #15010</a></li>
<li>Fix passthrough of atranscription into kwargs going to upstream provider - <a href="https://github.com/BerriAI/litellm/pull/15005" target="_blank" rel="noopener noreferrer">PR #15005</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/oci">OCI</a></strong>
<ul>
<li>Fix OCI Generative AI Integration when using Proxy - <a href="https://github.com/BerriAI/litellm/pull/15072" target="_blank" rel="noopener noreferrer">PR #15072</a></li>
</ul>
</li>
<li><strong>General</strong>
<ul>
<li>Fix: Authorization header to use correct "Bearer" capitalization - <a href="https://github.com/BerriAI/litellm/pull/14764" target="_blank" rel="noopener noreferrer">PR #14764</a></li>
<li>Bug fix: gpt-5-chat-latest has incorrect max_input_tokens value - <a href="https://github.com/BerriAI/litellm/pull/15116" target="_blank" rel="noopener noreferrer">PR #15116</a></li>
<li>Update request handling for original exceptions - <a href="https://github.com/BerriAI/litellm/pull/15013" target="_blank" rel="noopener noreferrer">PR #15013</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-provider-support">New Provider Support<a href="https://docs.litellm.ai/release_notes/v1-77-7#new-provider-support" class="hash-link" aria-label="Direct link to New Provider Support" title="Direct link to New Provider Support">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/providers/lemonade">AMD Lemonade</a></strong>
<ul>
<li>Add AMD Lemonade provider support - <a href="https://github.com/BerriAI/litellm/pull/14840" target="_blank" rel="noopener noreferrer">PR #14840</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-77-7#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-77-7#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/response_api">Responses API</a></strong></p>
<ul>
<li>Return Cost for Responses API Streaming requests - <a href="https://github.com/BerriAI/litellm/pull/15053" target="_blank" rel="noopener noreferrer">PR #15053</a></li>
</ul>
</li>
<li>
<p><strong><a href="https://docs.litellm.ai/docs/providers/gemini">/generateContent</a></strong></p>
<ul>
<li>Add full support for native Gemini API translation - <a href="https://github.com/BerriAI/litellm/pull/15029" target="_blank" rel="noopener noreferrer">PR #15029</a></li>
</ul>
</li>
<li>
<p><strong>Passthrough Gemini Routes</strong></p>
<ul>
<li>Add Gemini generateContent passthrough cost tracking - <a href="https://github.com/BerriAI/litellm/pull/15014" target="_blank" rel="noopener noreferrer">PR #15014</a></li>
<li>Add streamGenerateContent cost tracking in passthrough - <a href="https://github.com/BerriAI/litellm/pull/15199" target="_blank" rel="noopener noreferrer">PR #15199</a></li>
</ul>
</li>
<li>
<p><strong>Passthrough Vertex AI Routes</strong></p>
<ul>
<li>Add cost tracking for Vertex AI Passthrough <code>/predict</code> endpoint - <a href="https://github.com/BerriAI/litellm/pull/15019" target="_blank" rel="noopener noreferrer">PR #15019</a></li>
<li>Add cost tracking for Vertex AI Live API WebSocket Passthrough - <a href="https://github.com/BerriAI/litellm/pull/14956" target="_blank" rel="noopener noreferrer">PR #14956</a></li>
</ul>
</li>
<li>
<p><strong>General</strong></p>
<ul>
<li>Preserve Whitespace Characters in Model Response Streams - <a href="https://github.com/BerriAI/litellm/pull/15160" target="_blank" rel="noopener noreferrer">PR #15160</a></li>
<li>Add provider name to payload specification - <a href="https://github.com/BerriAI/litellm/pull/15130" target="_blank" rel="noopener noreferrer">PR #15130</a></li>
<li>Ensure query params are forwarded from origin url to downstream request - <a href="https://github.com/BerriAI/litellm/pull/15087" target="_blank" rel="noopener noreferrer">PR #15087</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-77-7#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-77-7#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li>
<p><strong>Virtual Keys</strong></p>
<ul>
<li>Ensure LLM_API_KEYs can access pass through routes - <a href="https://github.com/BerriAI/litellm/pull/15115" target="_blank" rel="noopener noreferrer">PR #15115</a></li>
<li>Support 'guaranteed_throughput' when setting limits on keys belonging to a team - <a href="https://github.com/BerriAI/litellm/pull/15120" target="_blank" rel="noopener noreferrer">PR #15120</a></li>
</ul>
</li>
<li>
<p><strong>Models + Endpoints</strong></p>
<ul>
<li>Ensure OCI secret fields not shared on /models and /v1/models endpoints - <a href="https://github.com/BerriAI/litellm/pull/15085" target="_blank" rel="noopener noreferrer">PR #15085</a></li>
<li>Add snowflake on UI - <a href="https://github.com/BerriAI/litellm/pull/15083" target="_blank" rel="noopener noreferrer">PR #15083</a></li>
<li>Make UI theme settings publicly accessible for custom branding - <a href="https://github.com/BerriAI/litellm/pull/15074" target="_blank" rel="noopener noreferrer">PR #15074</a></li>
</ul>
</li>
<li>
<p><strong>Admin Settings</strong></p>
<ul>
<li>Ensure OTEL settings are saved in DB after set on UI - <a href="https://github.com/BerriAI/litellm/pull/15118" target="_blank" rel="noopener noreferrer">PR #15118</a></li>
<li>Top api key tags - <a href="https://github.com/BerriAI/litellm/pull/15151" target="_blank" rel="noopener noreferrer">PR #15151</a>, <a href="https://github.com/BerriAI/litellm/pull/15156" target="_blank" rel="noopener noreferrer">PR #15156</a></li>
</ul>
</li>
<li>
<p><strong>MCP</strong></p>
<ul>
<li>show health status of MCP servers - <a href="https://github.com/BerriAI/litellm/pull/15185" target="_blank" rel="noopener noreferrer">PR #15185</a></li>
<li>allow setting extra headers on the UI - <a href="https://github.com/BerriAI/litellm/pull/15185" target="_blank" rel="noopener noreferrer">PR #15185</a></li>
<li>allow editing allowed tools on the UI - <a href="https://github.com/BerriAI/litellm/pull/15185" target="_blank" rel="noopener noreferrer">PR #15185</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes-1">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-77-7#bug-fixes-1" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li>
<p><strong>Virtual Keys</strong></p>
<ul>
<li>(security) prevent user key from updating other user keys - <a href="https://github.com/BerriAI/litellm/pull/15201" target="_blank" rel="noopener noreferrer">PR #15201</a></li>
<li>(security) don't return all keys with blank key alias on /v2/key/info - <a href="https://github.com/BerriAI/litellm/pull/15201" target="_blank" rel="noopener noreferrer">PR #15201</a></li>
<li>Fix Session Token Cookie Infinite Logout Loop - <a href="https://github.com/BerriAI/litellm/pull/15146" target="_blank" rel="noopener noreferrer">PR #15146</a></li>
</ul>
</li>
<li>
<p><strong>Models + Endpoints</strong></p>
<ul>
<li>Make UI theme settings publicly accessible for custom branding - <a href="https://github.com/BerriAI/litellm/pull/15074" target="_blank" rel="noopener noreferrer">PR #15074</a></li>
</ul>
</li>
<li>
<p><strong>Teams</strong></p>
<ul>
<li>fix failed copy to clipboard for http ui - <a href="https://github.com/BerriAI/litellm/pull/15195" target="_blank" rel="noopener noreferrer">PR #15195</a></li>
</ul>
</li>
<li>
<p><strong>Logs</strong></p>
<ul>
<li>fix logs page render logs on filter lookup - <a href="https://github.com/BerriAI/litellm/pull/15195" target="_blank" rel="noopener noreferrer">PR #15195</a></li>
<li>fix lookup list of end users (migrate to more efficient /customers/list lookup) - <a href="https://github.com/BerriAI/litellm/pull/15195" target="_blank" rel="noopener noreferrer">PR #15195</a></li>
</ul>
</li>
<li>
<p><strong>Test key</strong></p>
<ul>
<li>update selected model on key change - <a href="https://github.com/BerriAI/litellm/pull/15197" target="_blank" rel="noopener noreferrer">PR #15197</a></li>
</ul>
</li>
<li>
<p><strong>Dashboard</strong></p>
<ul>
<li>Fix LiteLLM model name fallback in dashboard overview - <a href="https://github.com/BerriAI/litellm/pull/14998" target="_blank" rel="noopener noreferrer">PR #14998</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="logging--guardrail--prompt-management-integrations">Logging / Guardrail / Prompt Management Integrations<a href="https://docs.litellm.ai/release_notes/v1-77-7#logging--guardrail--prompt-management-integrations" class="hash-link" aria-label="Direct link to Logging / Guardrail / Prompt Management Integrations" title="Direct link to Logging / Guardrail / Prompt Management Integrations">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-3">Features<a href="https://docs.litellm.ai/release_notes/v1-77-7#features-3" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/observability/otel">OpenTelemetry</a></strong>
<ul>
<li>Use generation_name for span naming in logging method - <a href="https://github.com/BerriAI/litellm/pull/14799" target="_blank" rel="noopener noreferrer">PR #14799</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging#langfuse">Langfuse</a></strong>
<ul>
<li>Handle non-serializable objects in Langfuse logging - <a href="https://github.com/BerriAI/litellm/pull/15148" target="_blank" rel="noopener noreferrer">PR #15148</a></li>
<li>Set usage_details.total in langfuse integration - <a href="https://github.com/BerriAI/litellm/pull/15015" target="_blank" rel="noopener noreferrer">PR #15015</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/prometheus">Prometheus</a></strong>
<ul>
<li>support custom metadata labels on key/team - <a href="https://github.com/BerriAI/litellm/pull/15094" target="_blank" rel="noopener noreferrer">PR #15094</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails">Guardrails<a href="https://docs.litellm.ai/release_notes/v1-77-7#guardrails" class="hash-link" aria-label="Direct link to Guardrails" title="Direct link to Guardrails">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/guardrails">Javelin</a></strong>
<ul>
<li>Add Javelin standalone guardrails integration for LiteLLM Proxy - <a href="https://github.com/BerriAI/litellm/pull/14983" target="_blank" rel="noopener noreferrer">PR #14983</a></li>
<li>Add logging for important status fields in guardrails - <a href="https://github.com/BerriAI/litellm/pull/15090" target="_blank" rel="noopener noreferrer">PR #15090</a></li>
<li>Don't run post_call guardrail if no text returned from Bedrock - <a href="https://github.com/BerriAI/litellm/pull/15106" target="_blank" rel="noopener noreferrer">PR #15106</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="prompt-management">Prompt Management<a href="https://docs.litellm.ai/release_notes/v1-77-7#prompt-management" class="hash-link" aria-label="Direct link to Prompt Management" title="Direct link to Prompt Management">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/prompt_management">GitLab</a></strong>
<ul>
<li>GitLab based Prompt manager - <a href="https://github.com/BerriAI/litellm/pull/14988" target="_blank" rel="noopener noreferrer">PR #14988</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-77-7#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li><strong>Cost Tracking</strong>
<ul>
<li>Proxy: end user cost tracking in the responses API - <a href="https://github.com/BerriAI/litellm/pull/15124" target="_blank" rel="noopener noreferrer">PR #15124</a></li>
</ul>
</li>
<li><strong>Parallel Request Limiter v3</strong>
<ul>
<li>Use well known redis cluster hashing algorithm - <a href="https://github.com/BerriAI/litellm/pull/15052" target="_blank" rel="noopener noreferrer">PR #15052</a></li>
<li>Fixes to dynamic rate limiter v3 - add saturation detection - <a href="https://github.com/BerriAI/litellm/pull/15119" target="_blank" rel="noopener noreferrer">PR #15119</a></li>
<li>Dynamic Rate Limiter v3 - fixes for detecting saturation + fixes for post saturation behavior - <a href="https://github.com/BerriAI/litellm/pull/15192" target="_blank" rel="noopener noreferrer">PR #15192</a></li>
</ul>
</li>
<li><strong>Teams</strong>
<ul>
<li>Add model specific tpm/rpm limits to teams on LiteLLM - <a href="https://github.com/BerriAI/litellm/pull/15044" target="_blank" rel="noopener noreferrer">PR #15044</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-gateway">MCP Gateway<a href="https://docs.litellm.ai/release_notes/v1-77-7#mcp-gateway" class="hash-link" aria-label="Direct link to MCP Gateway" title="Direct link to MCP Gateway">​</a></h2>
<ul>
<li><strong>Server Configuration</strong>
<ul>
<li>Specify forwardable headers, specify allowed/disallowed tools for MCP servers - <a href="https://github.com/BerriAI/litellm/pull/15002" target="_blank" rel="noopener noreferrer">PR #15002</a></li>
<li>Enforce server permissions on call tools - <a href="https://github.com/BerriAI/litellm/pull/15044" target="_blank" rel="noopener noreferrer">PR #15044</a></li>
<li>MCP Gateway Fine-grained Tools Addition - <a href="https://github.com/BerriAI/litellm/pull/15153" target="_blank" rel="noopener noreferrer">PR #15153</a></li>
</ul>
</li>
<li><strong>Bug Fixes</strong>
<ul>
<li>Remove servername prefix mcp tools tests - <a href="https://github.com/BerriAI/litellm/pull/14986" target="_blank" rel="noopener noreferrer">PR #14986</a></li>
<li>Resolve regression with duplicate Mcp-Protocol-Version header - <a href="https://github.com/BerriAI/litellm/pull/15050" target="_blank" rel="noopener noreferrer">PR #15050</a></li>
<li>Fix test_mcp_server.py - <a href="https://github.com/BerriAI/litellm/pull/15183" target="_blank" rel="noopener noreferrer">PR #15183</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements">Performance / Loadbalancing / Reliability improvements<a href="https://docs.litellm.ai/release_notes/v1-77-7#performance--loadbalancing--reliability-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements" title="Direct link to Performance / Loadbalancing / Reliability improvements">​</a></h2>
<ul>
<li><strong>Router Optimizations</strong>
<ul>
<li><strong>+62.5% P99 Latency Improvement</strong> - Remove router inefficiencies (from O(M*N) to O(1)) - <a href="https://github.com/BerriAI/litellm/pull/15046" target="_blank" rel="noopener noreferrer">PR #15046</a></li>
<li>Remove hasattr checks in Router - <a href="https://github.com/BerriAI/litellm/pull/15082" target="_blank" rel="noopener noreferrer">PR #15082</a></li>
<li>Remove Double Lookups - <a href="https://github.com/BerriAI/litellm/pull/15084" target="_blank" rel="noopener noreferrer">PR #15084</a></li>
<li>Optimize _filter_cooldown_deployments from O(n×m + k×n) to O(n) - <a href="https://github.com/BerriAI/litellm/pull/15091" target="_blank" rel="noopener noreferrer">PR #15091</a></li>
<li>Optimize unhealthy deployment filtering in retry path (O(n*m) → O(n+m)) - <a href="https://github.com/BerriAI/litellm/pull/15110" target="_blank" rel="noopener noreferrer">PR #15110</a></li>
</ul>
</li>
<li><strong>Cache Optimizations</strong>
<ul>
<li>Reduce complexity of InMemoryCache.evict_cache from O(n*log(n)) to O(log(n)) - <a href="https://github.com/BerriAI/litellm/pull/15000" target="_blank" rel="noopener noreferrer">PR #15000</a></li>
<li>Avoiding expensive operations when cache isn't available - <a href="https://github.com/BerriAI/litellm/pull/15182" target="_blank" rel="noopener noreferrer">PR #15182</a></li>
</ul>
</li>
<li><strong>Worker Management</strong>
<ul>
<li>Add proxy CLI option to recycle workers after N requests - <a href="https://github.com/BerriAI/litellm/pull/15007" target="_blank" rel="noopener noreferrer">PR #15007</a></li>
</ul>
</li>
<li><strong>Metrics &amp; Monitoring</strong>
<ul>
<li>LiteLLM Overhead metric tracking - Add support for tracking litellm overhead on cache hits - <a href="https://github.com/BerriAI/litellm/pull/15045" target="_blank" rel="noopener noreferrer">PR #15045</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates">Documentation Updates<a href="https://docs.litellm.ai/release_notes/v1-77-7#documentation-updates" class="hash-link" aria-label="Direct link to Documentation Updates" title="Direct link to Documentation Updates">​</a></h2>
<ul>
<li><strong>Provider Documentation</strong>
<ul>
<li>Update litellm docs from latest release - <a href="https://github.com/BerriAI/litellm/pull/15004" target="_blank" rel="noopener noreferrer">PR #15004</a></li>
<li>Add missing api_key parameter - <a href="https://github.com/BerriAI/litellm/pull/15058" target="_blank" rel="noopener noreferrer">PR #15058</a></li>
</ul>
</li>
<li><strong>General Documentation</strong>
<ul>
<li>Use docker compose instead of docker-compose - <a href="https://github.com/BerriAI/litellm/pull/15024" target="_blank" rel="noopener noreferrer">PR #15024</a></li>
<li>Add railtracks to projects that are using litellm - <a href="https://github.com/BerriAI/litellm/pull/15144" target="_blank" rel="noopener noreferrer">PR #15144</a></li>
<li>Perf: Last week improvement - <a href="https://github.com/BerriAI/litellm/pull/15193" target="_blank" rel="noopener noreferrer">PR #15193</a></li>
<li>Sync models GitHub documentation with Loom video and cross-reference - <a href="https://github.com/BerriAI/litellm/pull/15191" target="_blank" rel="noopener noreferrer">PR #15191</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="security-fixes">Security Fixes<a href="https://docs.litellm.ai/release_notes/v1-77-7#security-fixes" class="hash-link" aria-label="Direct link to Security Fixes" title="Direct link to Security Fixes">​</a></h2>
<ul>
<li><strong>JWT Token Security</strong> - Don't log JWT SSO token on .info() log - <a href="https://github.com/BerriAI/litellm/pull/15145" target="_blank" rel="noopener noreferrer">PR #15145</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-77-7#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@herve-ves made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14998" target="_blank" rel="noopener noreferrer">PR #14998</a></li>
<li>@wenxi-onyx made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15008" target="_blank" rel="noopener noreferrer">PR #15008</a></li>
<li>@jpetrucciani made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15005" target="_blank" rel="noopener noreferrer">PR #15005</a></li>
<li>@abhijitjavelin made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14983" target="_blank" rel="noopener noreferrer">PR #14983</a></li>
<li>@ZeroClover made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15039" target="_blank" rel="noopener noreferrer">PR #15039</a></li>
<li>@cedarm made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15043" target="_blank" rel="noopener noreferrer">PR #15043</a></li>
<li>@Isydmr made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15025" target="_blank" rel="noopener noreferrer">PR #15025</a></li>
<li>@serializer made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15013" target="_blank" rel="noopener noreferrer">PR #15013</a></li>
<li>@eddierichter-amd made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14840" target="_blank" rel="noopener noreferrer">PR #14840</a></li>
<li>@malags made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15000" target="_blank" rel="noopener noreferrer">PR #15000</a></li>
<li>@henryhwang made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15029" target="_blank" rel="noopener noreferrer">PR #15029</a></li>
<li>@plafleur made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15111" target="_blank" rel="noopener noreferrer">PR #15111</a></li>
<li>@tyler-liner made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14799" target="_blank" rel="noopener noreferrer">PR #14799</a></li>
<li>@Amir-R25 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15144" target="_blank" rel="noopener noreferrer">PR #15144</a></li>
<li>@georg-wolflein made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15124" target="_blank" rel="noopener noreferrer">PR #15124</a></li>
<li>@niharm made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15140" target="_blank" rel="noopener noreferrer">PR #15140</a></li>
<li>@anthony-liner made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15015" target="_blank" rel="noopener noreferrer">PR #15015</a></li>
<li>@rishiganesh2002 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15153" target="_blank" rel="noopener noreferrer">PR #15153</a></li>
<li>@danielaskdd made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15160" target="_blank" rel="noopener noreferrer">PR #15160</a></li>
<li>@JVenberg made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15146" target="_blank" rel="noopener noreferrer">PR #15146</a></li>
<li>@speglich made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/15072" target="_blank" rel="noopener noreferrer">PR #15072</a></li>
<li>@daily-kim made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14764" target="_blank" rel="noopener noreferrer">PR #14764</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog"><strong><a href="https://github.com/BerriAI/litellm/compare/v1.77.5.rc.4...v1.77.7.rc.1" target="_blank" rel="noopener noreferrer">Full Changelog</a></strong><a href="https://docs.litellm.ai/release_notes/v1-77-7#full-changelog" class="hash-link" aria-label="Direct link to full-changelog" title="Direct link to full-changelog">​</a></h2>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
    <entry>
        <title type="html"><![CDATA[v1.77.5-stable - MCP OAuth 2.0 Support]]></title>
        <id>https://docs.litellm.ai/release_notes/v1-77-5</id>
        <link href="https://docs.litellm.ai/release_notes/v1-77-5"/>
        <updated>2025-09-29T10:00:00.000Z</updated>
        <summary type="html"><![CDATA[Deploy this version]]></summary>
        <content type="html"><![CDATA[<h2 class="anchor anchorWithStickyNavbar_LWe7" id="deploy-this-version">Deploy this version<a href="https://docs.litellm.ai/release_notes/v1-77-5#deploy-this-version" class="hash-link" aria-label="Direct link to Deploy this version" title="Direct link to Deploy this version">​</a></h2>
<div class="tabs-container tabList__CuJ"><ul role="tablist" aria-orientation="horizontal" class="tabs"><li role="tab" tabindex="0" aria-selected="true" class="tabs__item tabItem_LNqP tabs__item--active">Docker</li><li role="tab" tabindex="-1" aria-selected="false" class="tabs__item tabItem_LNqP">Pip</li></ul><div class="margin-top--md"><div role="tabpanel" class="tabItem_Ymn6"><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">docker run litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">docker run \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-e STORE_MODEL_IN_DB=True \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">-p 4000:4000 \</span><br></span><span class="token-line" style="color:#393A34"><span class="token plain">docker.litellm.ai/berriai/litellm:v1.77.5-stable</span><br></span></code></pre></div></div></div><div role="tabpanel" class="tabItem_Ymn6" hidden=""><div class="language-showLineNumbers language-showlinenumbers codeBlockContainer_Ckt0 theme-code-block" style="--prism-color:#393A34;--prism-background-color:#f6f8fa"><div class="codeBlockTitle_OeMC">pip install litellm</div><div class="codeBlockContent_QJqH"><pre tabindex="0" class="prism-code language-showlinenumbers codeBlock_bY9V thin-scrollbar" style="color:#393A34;background-color:#f6f8fa"><code class="codeBlockLines_e6Vv"><span class="token-line" style="color:#393A34"><span class="token plain">pip install litellm==1.77.5</span><br></span></code></pre></div></div></div></div></div>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="key-highlights">Key Highlights<a href="https://docs.litellm.ai/release_notes/v1-77-5#key-highlights" class="hash-link" aria-label="Direct link to Key Highlights" title="Direct link to Key Highlights">​</a></h2>
<ul>
<li><strong>MCP OAuth 2.0 Support</strong> - Enhanced authentication for Model Context Protocol integrations</li>
<li><strong>Scheduled Key Rotations</strong> - Automated key rotation capabilities for enhanced security</li>
<li><strong>New Gemini 2.5 Flash &amp; Flash-lite Models</strong> - Latest September 2025 preview models with improved pricing and features</li>
<li><strong>Performance Improvements</strong> - 54% RPS improvement</li>
</ul>
<hr>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="performance-improvements---54-rps-improvement">Performance Improvements - 54% RPS Improvement<a href="https://docs.litellm.ai/release_notes/v1-77-5#performance-improvements---54-rps-improvement" class="hash-link" aria-label="Direct link to Performance Improvements - 54% RPS Improvement" title="Direct link to Performance Improvements - 54% RPS Improvement">​</a></h3>
<div style="background-size:cover;background-repeat:no-repeat;position:relative;background-image:url(&quot;data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAFCAYAAAB8ZH1oAAAACXBIWXMAACxLAAAsSwGlPZapAAAAsUlEQVR4nB3GzWrCQBRA4Xl7u6gu3FsK4iKLPkChgnTtUop0UWOqmSZkUvM7zr2Jp1QOHxxzOlVibS6udNK1naBId+3lcq3u/58PXkxR1JzTlDJ3WGeJyhcWbslzueKpWPLVH2AAoyrchpFeex5/5qx/N7jGUXUXds0Hk3TKvv3E+OBhhFUc8Zq9QYDvOCHNLHVTc2wTHvYzTAhBJYjOtnPN6lwHGVRE7nzwiqKbw7v+AcUMt4bRbIoWAAAAAElFTkSuQmCC&quot;)"><svg style="width:100%;height:auto;max-width:100%;margin-bottom:-4px" width="640" height="332"></svg><noscript><img style="width:100%;height:auto;max-width:100%;margin-bottom:-4px;position:absolute;top:0;left:0" src="/assets/ideal-img/perf_77_5.90cb032.640.png" srcset="/assets/ideal-img/perf_77_5.90cb032.640.png 640w,/assets/ideal-img/perf_77_5.2a1a598.1920.png 1920w" width="640" height="332"></noscript></div>
<br>
<p>This release brings a 54% RPS improvement (1,040 → 1,602 RPS, aggregated) per instance.</p>
<p>The improvement comes from fixing O(n²) inefficiencies in the LiteLLM Router, primarily caused by repeated use of <code>in</code> statements inside loops over large arrays.</p>
<p>Tests were run with a database-only setup (no cache hits).</p>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="test-setup">Test Setup<a href="https://docs.litellm.ai/release_notes/v1-77-5#test-setup" class="hash-link" aria-label="Direct link to Test Setup" title="Direct link to Test Setup">​</a></h4>
<p>All benchmarks were executed using Locust with 1,000 concurrent users and a ramp-up of 500. The environment was configured to stress the routing layer and eliminate caching as a variable.</p>
<p><strong>System Specs</strong></p>
<ul>
<li><strong>CPU:</strong> 8 vCPUs</li>
<li><strong>Memory:</strong> 32 GB RAM</li>
</ul>
<p><strong>Configuration (config.yaml)</strong></p>
<p>View the complete configuration: <a href="https://gist.github.com/AlexsanderHamir/53f7d554a5d2afcf2c4edb5b6be68ff4" target="_blank" rel="noopener noreferrer">gist.github.com/AlexsanderHamir/config.yaml</a></p>
<p><strong>Load Script (no_cache_hits.py)</strong></p>
<p>View the complete load testing script: <a href="https://gist.github.com/AlexsanderHamir/42c33d7a4dc7a57f56a78b560dee3a42" target="_blank" rel="noopener noreferrer">gist.github.com/AlexsanderHamir/no_cache_hits.py</a></p>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-models--updated-models">New Models / Updated Models<a href="https://docs.litellm.ai/release_notes/v1-77-5#new-models--updated-models" class="hash-link" aria-label="Direct link to New Models / Updated Models" title="Direct link to New Models / Updated Models">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-model-support">New Model Support<a href="https://docs.litellm.ai/release_notes/v1-77-5#new-model-support" class="hash-link" aria-label="Direct link to New Model Support" title="Direct link to New Model Support">​</a></h4>
<table><thead><tr><th>Provider</th><th>Model</th><th>Context Window</th><th>Input ($/1M tokens)</th><th>Output ($/1M tokens)</th><th>Features</th></tr></thead><tbody><tr><td>Gemini</td><td><code>gemini-2.5-flash-preview-09-2025</code></td><td>1M</td><td>$0.30</td><td>$2.50</td><td>Chat, reasoning, vision, audio</td></tr><tr><td>Gemini</td><td><code>gemini-2.5-flash-lite-preview-09-2025</code></td><td>1M</td><td>$0.10</td><td>$0.40</td><td>Chat, reasoning, vision, audio</td></tr><tr><td>Gemini</td><td><code>gemini-flash-latest</code></td><td>1M</td><td>$0.30</td><td>$2.50</td><td>Chat, reasoning, vision, audio</td></tr><tr><td>Gemini</td><td><code>gemini-flash-lite-latest</code></td><td>1M</td><td>$0.10</td><td>$0.40</td><td>Chat, reasoning, vision, audio</td></tr><tr><td>DeepSeek</td><td><code>deepseek-chat</code></td><td>131K</td><td>$0.60</td><td>$1.70</td><td>Chat, function calling, caching</td></tr><tr><td>DeepSeek</td><td><code>deepseek-reasoner</code></td><td>131K</td><td>$0.60</td><td>$1.70</td><td>Chat, reasoning</td></tr><tr><td>Bedrock</td><td><code>deepseek.v3-v1:0</code></td><td>164K</td><td>$0.58</td><td>$1.68</td><td>Chat, reasoning, function calling</td></tr><tr><td>Azure</td><td><code>azure/gpt-5-codex</code></td><td>272K</td><td>$1.25</td><td>$10.00</td><td>Responses API, reasoning, vision</td></tr><tr><td>OpenAI</td><td><code>gpt-5-codex</code></td><td>272K</td><td>$1.25</td><td>$10.00</td><td>Responses API, reasoning, vision</td></tr><tr><td>SambaNova</td><td><code>sambanova/DeepSeek-V3.1</code></td><td>33K</td><td>$3.00</td><td>$4.50</td><td>Chat, reasoning, function calling</td></tr><tr><td>SambaNova</td><td><code>sambanova/gpt-oss-120b</code></td><td>131K</td><td>$3.00</td><td>$4.50</td><td>Chat, reasoning, function calling</td></tr><tr><td>Bedrock</td><td><code>qwen.qwen3-coder-480b-a35b-v1:0</code></td><td>262K</td><td>$0.22</td><td>$1.80</td><td>Chat, reasoning, function calling</td></tr><tr><td>Bedrock</td><td><code>qwen.qwen3-235b-a22b-2507-v1:0</code></td><td>262K</td><td>$0.22</td><td>$0.88</td><td>Chat, reasoning, function calling</td></tr><tr><td>Bedrock</td><td><code>qwen.qwen3-coder-30b-a3b-v1:0</code></td><td>262K</td><td>$0.15</td><td>$0.60</td><td>Chat, reasoning, function calling</td></tr><tr><td>Bedrock</td><td><code>qwen.qwen3-32b-v1:0</code></td><td>131K</td><td>$0.15</td><td>$0.60</td><td>Chat, reasoning, function calling</td></tr><tr><td>Vertex AI</td><td><code>vertex_ai/qwen/qwen3-next-80b-a3b-instruct-maas</code></td><td>262K</td><td>$0.15</td><td>$1.20</td><td>Chat, function calling</td></tr><tr><td>Vertex AI</td><td><code>vertex_ai/qwen/qwen3-next-80b-a3b-thinking-maas</code></td><td>262K</td><td>$0.15</td><td>$1.20</td><td>Chat, function calling</td></tr><tr><td>Vertex AI</td><td><code>vertex_ai/deepseek-ai/deepseek-v3.1-maas</code></td><td>164K</td><td>$1.35</td><td>$5.40</td><td>Chat, reasoning, function calling</td></tr><tr><td>OpenRouter</td><td><code>openrouter/x-ai/grok-4-fast:free</code></td><td>2M</td><td>$0.00</td><td>$0.00</td><td>Chat, reasoning, function calling</td></tr><tr><td>XAI</td><td><code>xai/grok-4-fast-reasoning</code></td><td>2M</td><td>$0.20</td><td>$0.50</td><td>Chat, reasoning, function calling</td></tr><tr><td>XAI</td><td><code>xai/grok-4-fast-non-reasoning</code></td><td>2M</td><td>$0.20</td><td>$0.50</td><td>Chat, function calling</td></tr></tbody></table>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features">Features<a href="https://docs.litellm.ai/release_notes/v1-77-5#features" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/providers/gemini">Gemini</a></strong>
<ul>
<li>Added Gemini 2.5 Flash and Flash-lite preview models (September 2025 release) with improved pricing - <a href="https://github.com/BerriAI/litellm/pull/14948" target="_blank" rel="noopener noreferrer">PR #14948</a></li>
<li>Added new Anthropic web fetch tool support - <a href="https://github.com/BerriAI/litellm/pull/14951" target="_blank" rel="noopener noreferrer">PR #14951</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/xai">XAI</a></strong>
<ul>
<li>Add xai/grok-4-fast models - <a href="https://github.com/BerriAI/litellm/pull/14833" target="_blank" rel="noopener noreferrer">PR #14833</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong>
<ul>
<li>Updated Claude Sonnet 4 configs to reflect million-token context window pricing - <a href="https://github.com/BerriAI/litellm/pull/14639" target="_blank" rel="noopener noreferrer">PR #14639</a></li>
<li>Added supported text field to anthropic citation response - <a href="https://github.com/BerriAI/litellm/pull/14164" target="_blank" rel="noopener noreferrer">PR #14164</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/bedrock">Bedrock</a></strong>
<ul>
<li>Added support for Qwen models family &amp; Deepseek 3.1 to Amazon Bedrock - <a href="https://github.com/BerriAI/litellm/pull/14845" target="_blank" rel="noopener noreferrer">PR #14845</a></li>
<li>Support requestMetadata in Bedrock Converse API - <a href="https://github.com/BerriAI/litellm/pull/14570" target="_blank" rel="noopener noreferrer">PR #14570</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/vertex">Vertex AI</a></strong>
<ul>
<li>Added vertex_ai/qwen models and azure/gpt-5-codex - <a href="https://github.com/BerriAI/litellm/pull/14844" target="_blank" rel="noopener noreferrer">PR #14844</a></li>
<li>Update vertex ai qwen model pricing - <a href="https://github.com/BerriAI/litellm/pull/14828" target="_blank" rel="noopener noreferrer">PR #14828</a></li>
<li>Vertex AI Context Caching: use Vertex ai API v1 instead of v1beta1 and accept 'cachedContent' param - <a href="https://github.com/BerriAI/litellm/pull/14831" target="_blank" rel="noopener noreferrer">PR #14831</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/sambanova">SambaNova</a></strong>
<ul>
<li>Add sambanova deepseek v3.1 and gpt-oss-120b - <a href="https://github.com/BerriAI/litellm/pull/14866" target="_blank" rel="noopener noreferrer">PR #14866</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI</a></strong>
<ul>
<li>Fix inconsistent token configs for gpt-5 models - <a href="https://github.com/BerriAI/litellm/pull/14942" target="_blank" rel="noopener noreferrer">PR #14942</a></li>
<li>GPT-3.5-Turbo price updated - <a href="https://github.com/BerriAI/litellm/pull/14858" target="_blank" rel="noopener noreferrer">PR #14858</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/openrouter">OpenRouter</a></strong>
<ul>
<li>Add gpt-5 and gpt-5-codex to OpenRouter cost map - <a href="https://github.com/BerriAI/litellm/pull/14879" target="_blank" rel="noopener noreferrer">PR #14879</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/vllm">VLLM</a></strong>
<ul>
<li>Fix vllm passthrough - <a href="https://github.com/BerriAI/litellm/pull/14778" target="_blank" rel="noopener noreferrer">PR #14778</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/image_generation">Flux</a></strong>
<ul>
<li>Support flux image edit - <a href="https://github.com/BerriAI/litellm/pull/14790" target="_blank" rel="noopener noreferrer">PR #14790</a></li>
</ul>
</li>
</ul>
<h3 class="anchor anchorWithStickyNavbar_LWe7" id="bug-fixes">Bug Fixes<a href="https://docs.litellm.ai/release_notes/v1-77-5#bug-fixes" class="hash-link" aria-label="Direct link to Bug Fixes" title="Direct link to Bug Fixes">​</a></h3>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/providers/anthropic">Anthropic</a></strong>
<ul>
<li>Fix: Support claude code auth via subscription (anthropic) - <a href="https://github.com/BerriAI/litellm/pull/14821" target="_blank" rel="noopener noreferrer">PR #14821</a></li>
<li>Fix Anthropic streaming IDs - <a href="https://github.com/BerriAI/litellm/pull/14965" target="_blank" rel="noopener noreferrer">PR #14965</a></li>
<li>Revert incorrect changes to sonnet-4 max output tokens - <a href="https://github.com/BerriAI/litellm/pull/14933" target="_blank" rel="noopener noreferrer">PR #14933</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/openai">OpenAI</a></strong>
<ul>
<li>Fix a bug where openai image edit silently ignores multiple images - <a href="https://github.com/BerriAI/litellm/pull/14893" target="_blank" rel="noopener noreferrer">PR #14893</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/providers/vllm">VLLM</a></strong>
<ul>
<li>Fix: vLLM provider's rerank endpoint from /v1/rerank to /rerank - <a href="https://github.com/BerriAI/litellm/pull/14938" target="_blank" rel="noopener noreferrer">PR #14938</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="new-provider-support">New Provider Support<a href="https://docs.litellm.ai/release_notes/v1-77-5#new-provider-support" class="hash-link" aria-label="Direct link to New Provider Support" title="Direct link to New Provider Support">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/providers/wandb">W&amp;B Inference</a></strong>
<ul>
<li>Add W&amp;B Inference to LiteLLM - <a href="https://github.com/BerriAI/litellm/pull/14416" target="_blank" rel="noopener noreferrer">PR #14416</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="llm-api-endpoints">LLM API Endpoints<a href="https://docs.litellm.ai/release_notes/v1-77-5#llm-api-endpoints" class="hash-link" aria-label="Direct link to LLM API Endpoints" title="Direct link to LLM API Endpoints">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-1">Features<a href="https://docs.litellm.ai/release_notes/v1-77-5#features-1" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li><strong>General</strong>
<ul>
<li>Add SDK support for additional headers - <a href="https://github.com/BerriAI/litellm/pull/14761" target="_blank" rel="noopener noreferrer">PR #14761</a></li>
<li>Add shared_session parameter for aiohttp ClientSession reuse - <a href="https://github.com/BerriAI/litellm/pull/14721" target="_blank" rel="noopener noreferrer">PR #14721</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs">Bugs<a href="https://docs.litellm.ai/release_notes/v1-77-5#bugs" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>General</strong>
<ul>
<li>Fix: Streaming tool call index assignment for multiple tool calls - <a href="https://github.com/BerriAI/litellm/pull/14587" target="_blank" rel="noopener noreferrer">PR #14587</a></li>
<li>Fix load credentials in token counter proxy - <a href="https://github.com/BerriAI/litellm/pull/14808" target="_blank" rel="noopener noreferrer">PR #14808</a></li>
</ul>
</li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="management-endpoints--ui">Management Endpoints / UI<a href="https://docs.litellm.ai/release_notes/v1-77-5#management-endpoints--ui" class="hash-link" aria-label="Direct link to Management Endpoints / UI" title="Direct link to Management Endpoints / UI">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-2">Features<a href="https://docs.litellm.ai/release_notes/v1-77-5#features-2" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li><strong>Proxy CLI Auth</strong>
<ul>
<li>Allow re-using cli auth token - <a href="https://github.com/BerriAI/litellm/pull/14780" target="_blank" rel="noopener noreferrer">PR #14780</a></li>
<li>Create a python method to login using litellm proxy - <a href="https://github.com/BerriAI/litellm/pull/14782" target="_blank" rel="noopener noreferrer">PR #14782</a></li>
<li>Fixes for LiteLLM Proxy CLI to Auth to Gateway - <a href="https://github.com/BerriAI/litellm/pull/14836" target="_blank" rel="noopener noreferrer">PR #14836</a></li>
</ul>
</li>
</ul>
<p><strong>Virtual Keys</strong></p>
<ul>
<li>Initial support for scheduled key rotations - <a href="https://github.com/BerriAI/litellm/pull/14877" target="_blank" rel="noopener noreferrer">PR #14877</a></li>
<li>Allow scheduling key rotations when creating virtual keys - <a href="https://github.com/BerriAI/litellm/pull/14960" target="_blank" rel="noopener noreferrer">PR #14960</a></li>
</ul>
<p><strong>Models + Endpoints</strong></p>
<ul>
<li>Fix: added Oracle to provider's list - <a href="https://github.com/BerriAI/litellm/pull/14835" target="_blank" rel="noopener noreferrer">PR #14835</a></li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="bugs-1">Bugs<a href="https://docs.litellm.ai/release_notes/v1-77-5#bugs-1" class="hash-link" aria-label="Direct link to Bugs" title="Direct link to Bugs">​</a></h4>
<ul>
<li><strong>SSO</strong> - Fix: SSO "Clear" button writes empty values instead of removing SSO config - <a href="https://github.com/BerriAI/litellm/pull/14826" target="_blank" rel="noopener noreferrer">PR #14826</a></li>
<li><strong>Admin Settings</strong> - Remove useful links from admin settings - <a href="https://github.com/BerriAI/litellm/pull/14918" target="_blank" rel="noopener noreferrer">PR #14918</a></li>
<li><strong>Management Routes</strong> - Add /user/list to management routes - <a href="https://github.com/BerriAI/litellm/pull/14868" target="_blank" rel="noopener noreferrer">PR #14868</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="logging--guardrail--prompt-management-integrations">Logging / Guardrail / Prompt Management Integrations<a href="https://docs.litellm.ai/release_notes/v1-77-5#logging--guardrail--prompt-management-integrations" class="hash-link" aria-label="Direct link to Logging / Guardrail / Prompt Management Integrations" title="Direct link to Logging / Guardrail / Prompt Management Integrations">​</a></h2>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="features-3">Features<a href="https://docs.litellm.ai/release_notes/v1-77-5#features-3" class="hash-link" aria-label="Direct link to Features" title="Direct link to Features">​</a></h4>
<ul>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging#datadog">DataDog</a></strong>
<ul>
<li>Logging - <code>datadog</code> callback Log message content w/o sending to datadog - <a href="https://github.com/BerriAI/litellm/pull/14909" target="_blank" rel="noopener noreferrer">PR #14909</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging#langfuse">Langfuse</a></strong>
<ul>
<li>Adding langfuse usage details for cached tokens - <a href="https://github.com/BerriAI/litellm/pull/10955" target="_blank" rel="noopener noreferrer">PR #10955</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging#opik">Opik</a></strong>
<ul>
<li>Improve opik integration code - <a href="https://github.com/BerriAI/litellm/pull/14888" target="_blank" rel="noopener noreferrer">PR #14888</a></li>
</ul>
</li>
<li><strong><a href="https://docs.litellm.ai/docs/proxy/logging#sqs">SQS</a></strong>
<ul>
<li>Error logging support for SQS Logger - <a href="https://github.com/BerriAI/litellm/pull/14974" target="_blank" rel="noopener noreferrer">PR #14974</a></li>
</ul>
</li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="guardrails">Guardrails<a href="https://docs.litellm.ai/release_notes/v1-77-5#guardrails" class="hash-link" aria-label="Direct link to Guardrails" title="Direct link to Guardrails">​</a></h4>
<ul>
<li><strong>LakeraAI v2 Guardrail</strong> - Ensure exception is raised correctly - <a href="https://github.com/BerriAI/litellm/pull/14867" target="_blank" rel="noopener noreferrer">PR #14867</a></li>
<li><strong>Presidio Guardrail</strong> - Support custom entity types in Presidio guardrail with Union[PiiEntityType, str] - <a href="https://github.com/BerriAI/litellm/pull/14899" target="_blank" rel="noopener noreferrer">PR #14899</a></li>
<li><strong>Noma Guardrail</strong> - Add noma guardrail provider to ui - <a href="https://github.com/BerriAI/litellm/pull/14415" target="_blank" rel="noopener noreferrer">PR #14415</a></li>
</ul>
<h4 class="anchor anchorWithStickyNavbar_LWe7" id="prompt-management">Prompt Management<a href="https://docs.litellm.ai/release_notes/v1-77-5#prompt-management" class="hash-link" aria-label="Direct link to Prompt Management" title="Direct link to Prompt Management">​</a></h4>
<ul>
<li><strong>BitBucket Integration</strong> - Add BitBucket Integration for Prompt Management - <a href="https://github.com/BerriAI/litellm/pull/14882" target="_blank" rel="noopener noreferrer">PR #14882</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="spend-tracking-budgets-and-rate-limiting">Spend Tracking, Budgets and Rate Limiting<a href="https://docs.litellm.ai/release_notes/v1-77-5#spend-tracking-budgets-and-rate-limiting" class="hash-link" aria-label="Direct link to Spend Tracking, Budgets and Rate Limiting" title="Direct link to Spend Tracking, Budgets and Rate Limiting">​</a></h2>
<ul>
<li><strong>Service Tier Pricing</strong> - Add service_tier based pricing support for openai (BOTH Service &amp; Priority Support) - <a href="https://github.com/BerriAI/litellm/pull/14796" target="_blank" rel="noopener noreferrer">PR #14796</a></li>
<li><strong>Cost Tracking</strong> - Show input, output, tool call cost breakdown in StandardLoggingPayload - <a href="https://github.com/BerriAI/litellm/pull/14921" target="_blank" rel="noopener noreferrer">PR #14921</a></li>
<li><strong>Parallel Request Limiter v3</strong>
<ul>
<li>Ensure Lua scripts can execute on redis cluster - <a href="https://github.com/BerriAI/litellm/pull/14968" target="_blank" rel="noopener noreferrer">PR #14968</a></li>
<li>Fix: get metadata info from both metadata and litellm_metadata fields - <a href="https://github.com/BerriAI/litellm/pull/14783" target="_blank" rel="noopener noreferrer">PR #14783</a></li>
</ul>
</li>
<li><strong>Priority Reservation</strong> - Fix: Priority Reservation: keys without priority metadata receive higher priority than keys with explicit priority configurations - <a href="https://github.com/BerriAI/litellm/pull/14832" target="_blank" rel="noopener noreferrer">PR #14832</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="mcp-gateway">MCP Gateway<a href="https://docs.litellm.ai/release_notes/v1-77-5#mcp-gateway" class="hash-link" aria-label="Direct link to MCP Gateway" title="Direct link to MCP Gateway">​</a></h2>
<ul>
<li><strong>MCP Configuration</strong> - Enable custom fields in mcp_info configuration - <a href="https://github.com/BerriAI/litellm/pull/14794" target="_blank" rel="noopener noreferrer">PR #14794</a></li>
<li><strong>MCP Tools</strong> - Remove server_name prefix from list_tools - <a href="https://github.com/BerriAI/litellm/pull/14720" target="_blank" rel="noopener noreferrer">PR #14720</a></li>
<li><strong>OAuth Flow</strong> - Initial commit for v2 oauth flow - <a href="https://github.com/BerriAI/litellm/pull/14964" target="_blank" rel="noopener noreferrer">PR #14964</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="performance--loadbalancing--reliability-improvements">Performance / Loadbalancing / Reliability improvements<a href="https://docs.litellm.ai/release_notes/v1-77-5#performance--loadbalancing--reliability-improvements" class="hash-link" aria-label="Direct link to Performance / Loadbalancing / Reliability improvements" title="Direct link to Performance / Loadbalancing / Reliability improvements">​</a></h2>
<ul>
<li><strong>Memory Leak Fix</strong> - Fix InMemoryCache unbounded growth when TTLs are set - <a href="https://github.com/BerriAI/litellm/pull/14869" target="_blank" rel="noopener noreferrer">PR #14869</a></li>
<li><strong>Cache Performance</strong> - Fix: cache root cause - <a href="https://github.com/BerriAI/litellm/pull/14827" target="_blank" rel="noopener noreferrer">PR #14827</a></li>
<li><strong>Concurrency Fix</strong> - Fix concurrency/scaling when many Python threads do streaming using <em>sync</em> completions - <a href="https://github.com/BerriAI/litellm/pull/14816" target="_blank" rel="noopener noreferrer">PR #14816</a></li>
<li><strong>Performance Optimization</strong> - Fix: reduce get_deployment cost to O(1) - <a href="https://github.com/BerriAI/litellm/pull/14967" target="_blank" rel="noopener noreferrer">PR #14967</a></li>
<li><strong>Performance Optimization</strong> - Fix: remove slow string operation - <a href="https://github.com/BerriAI/litellm/pull/14955" target="_blank" rel="noopener noreferrer">PR #14955</a></li>
<li><strong>DB Connection Management</strong> - Fix: DB connection state retries - <a href="https://github.com/BerriAI/litellm/pull/14925" target="_blank" rel="noopener noreferrer">PR #14925</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="documentation-updates">Documentation Updates<a href="https://docs.litellm.ai/release_notes/v1-77-5#documentation-updates" class="hash-link" aria-label="Direct link to Documentation Updates" title="Direct link to Documentation Updates">​</a></h2>
<ul>
<li><strong>Provider Documentation</strong> - Fix docs for provider_specific_params.md - <a href="https://github.com/BerriAI/litellm/pull/14787" target="_blank" rel="noopener noreferrer">PR #14787</a></li>
<li><strong>Model References</strong> - Update model references from gemini-pro to gemini-2.5-pro - <a href="https://github.com/BerriAI/litellm/pull/14775" target="_blank" rel="noopener noreferrer">PR #14775</a></li>
<li><strong>Letta Guide</strong> - Add Letta Guide documentation - <a href="https://github.com/BerriAI/litellm/pull/14798" target="_blank" rel="noopener noreferrer">PR #14798</a></li>
<li><strong>README</strong> - Make the README document clearer - <a href="https://github.com/BerriAI/litellm/pull/14860" target="_blank" rel="noopener noreferrer">PR #14860</a></li>
<li><strong>Session Management</strong> - Update docs for session management availability - <a href="https://github.com/BerriAI/litellm/pull/14914" target="_blank" rel="noopener noreferrer">PR #14914</a></li>
<li><strong>Cost Documentation</strong> - Add documentation for additional cost-related keys in custom pricing - <a href="https://github.com/BerriAI/litellm/pull/14949" target="_blank" rel="noopener noreferrer">PR #14949</a></li>
<li><strong>Azure Passthrough</strong> - Add azure passthrough documentation - <a href="https://github.com/BerriAI/litellm/pull/14958" target="_blank" rel="noopener noreferrer">PR #14958</a></li>
<li><strong>General Documentation</strong> - Doc updates sept 2025 - <a href="https://github.com/BerriAI/litellm/pull/14769" target="_blank" rel="noopener noreferrer">PR #14769</a>
<ul>
<li>Clarified bridging between endpoints and mode in docs.</li>
<li>Added Vertex AI Gemini API configuration as an alternative in relevant guides.
Linked AWS authentication info in the Bedrock guardrails documentation.</li>
<li>Added Cancel Response API usage with code snippets</li>
<li>Clarified that SSO (Single Sign-On) is free for up to 5 users:</li>
<li>Alphabetized sidebar, leaving quick start / intros at top of categories</li>
<li>Documented max_connections under cache_params.</li>
<li>Clarified IAM AssumeRole Policy requirements.</li>
<li>Added transform utilities example to Getting Started (showing request transformation).</li>
<li>Added references to models.litellm.ai as the full models list in various docs.</li>
<li>Added a code snippet for async_post_call_success_hook.</li>
<li>Removed broken links to callbacks management guide. - Reformatted and linked cookbooks + other relevant docs</li>
</ul>
</li>
<li><strong>Documentation Corrections</strong> - Corrected docs updates sept 2025 - <a href="https://github.com/BerriAI/litellm/pull/14916" target="_blank" rel="noopener noreferrer">PR #14916</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="new-contributors">New Contributors<a href="https://docs.litellm.ai/release_notes/v1-77-5#new-contributors" class="hash-link" aria-label="Direct link to New Contributors" title="Direct link to New Contributors">​</a></h2>
<ul>
<li>@uzaxirr made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14761" target="_blank" rel="noopener noreferrer">PR #14761</a></li>
<li>@xprilion made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14416" target="_blank" rel="noopener noreferrer">PR #14416</a></li>
<li>@CH-GAGANRAJ made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14779" target="_blank" rel="noopener noreferrer">PR #14779</a></li>
<li>@otaviofbrito made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14778" target="_blank" rel="noopener noreferrer">PR #14778</a></li>
<li>@danielmklein made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14639" target="_blank" rel="noopener noreferrer">PR #14639</a></li>
<li>@Jetemple made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14826" target="_blank" rel="noopener noreferrer">PR #14826</a></li>
<li>@akshoop made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14818" target="_blank" rel="noopener noreferrer">PR #14818</a></li>
<li>@hazyone made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14821" target="_blank" rel="noopener noreferrer">PR #14821</a></li>
<li>@leventov made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14816" target="_blank" rel="noopener noreferrer">PR #14816</a></li>
<li>@fabriciojoc made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/10955" target="_blank" rel="noopener noreferrer">PR #10955</a></li>
<li>@onlylonly made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14845" target="_blank" rel="noopener noreferrer">PR #14845</a></li>
<li>@Copilot made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14869" target="_blank" rel="noopener noreferrer">PR #14869</a></li>
<li>@arsh72 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14899" target="_blank" rel="noopener noreferrer">PR #14899</a></li>
<li>@berri-teddy made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14914" target="_blank" rel="noopener noreferrer">PR #14914</a></li>
<li>@vpbill made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14415" target="_blank" rel="noopener noreferrer">PR #14415</a></li>
<li>@kgritesh made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14893" target="_blank" rel="noopener noreferrer">PR #14893</a></li>
<li>@oytunkutrup1 made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14858" target="_blank" rel="noopener noreferrer">PR #14858</a></li>
<li>@nherment made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14933" target="_blank" rel="noopener noreferrer">PR #14933</a></li>
<li>@deepanshululla made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14974" target="_blank" rel="noopener noreferrer">PR #14974</a></li>
<li>@TeddyAmkie made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14758" target="_blank" rel="noopener noreferrer">PR #14758</a></li>
<li>@SmartManoj made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14775" target="_blank" rel="noopener noreferrer">PR #14775</a></li>
<li>@uc4w6c made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14720" target="_blank" rel="noopener noreferrer">PR #14720</a></li>
<li>@luizrennocosta made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14783" target="_blank" rel="noopener noreferrer">PR #14783</a></li>
<li>@AlexsanderHamir made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14827" target="_blank" rel="noopener noreferrer">PR #14827</a></li>
<li>@dharamendrak made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14721" target="_blank" rel="noopener noreferrer">PR #14721</a></li>
<li>@TomeHirata made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14164" target="_blank" rel="noopener noreferrer">PR #14164</a></li>
<li>@mrFranklin made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14860" target="_blank" rel="noopener noreferrer">PR #14860</a></li>
<li>@luisfucros made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14866" target="_blank" rel="noopener noreferrer">PR #14866</a></li>
<li>@huangyafei made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14879" target="_blank" rel="noopener noreferrer">PR #14879</a></li>
<li>@thiswillbeyourgithub made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14949" target="_blank" rel="noopener noreferrer">PR #14949</a></li>
<li>@Maximgitman made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14965" target="_blank" rel="noopener noreferrer">PR #14965</a></li>
<li>@subnet-dev made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14938" target="_blank" rel="noopener noreferrer">PR #14938</a></li>
<li>@22mSqRi made their first contribution in <a href="https://github.com/BerriAI/litellm/pull/14972" target="_blank" rel="noopener noreferrer">PR #14972</a></li>
</ul>
<hr>
<h2 class="anchor anchorWithStickyNavbar_LWe7" id="full-changelog"><strong><a href="https://github.com/BerriAI/litellm/compare/v1.77.3.rc.1...v1.77.5.rc.1" target="_blank" rel="noopener noreferrer">Full Changelog</a></strong><a href="https://docs.litellm.ai/release_notes/v1-77-5#full-changelog" class="hash-link" aria-label="Direct link to full-changelog" title="Direct link to full-changelog">​</a></h2>]]></content>
        <author>
            <name>Krrish Dholakia</name>
            <uri>https://www.linkedin.com/in/krish-d/</uri>
        </author>
        <author>
            <name>Ishaan Jaff</name>
            <uri>https://www.linkedin.com/in/reffajnaahsi/</uri>
        </author>
    </entry>
</feed>