--- title: "v1.72.0-stable" slug: "v1-72-0-stable" date: 2025-05-31T10:00:00 authors: - name: Krrish Dholakia title: CEO, LiteLLM url: https://www.linkedin.com/in/krish-d/ image_url: https://media.licdn.com/dms/image/v2/D4D03AQGrlsJ3aqpHmQ/profile-displayphoto-shrink_400_400/B4DZSAzgP7HYAg-/0/1737327772964?e=1749686400&v=beta&t=Hkl3U8Ps0VtvNxX0BNNq24b4dtX5wQaPFp6oiKCIHD8 - name: Ishaan Jaffer title: CTO, LiteLLM url: https://www.linkedin.com/in/reffajnaahsi/ image_url: https://pbs.twimg.com/profile_images/1613813310264340481/lz54oEiB_400x400.jpg hide_table_of_contents: false --- import Image from '@theme/IdealImage'; import Tabs from '@theme/Tabs'; import TabItem from '@theme/TabItem'; ## Deploy this version ``` showLineNumbers title="docker run litellm" docker run -e STORE_MODEL_IN_DB=True -p 4000:4000 ghcr.io/berriai/litellm:main-v1.72.0-stable ``` ``` showLineNumbers title="pip install litellm" pip install litellm==1.72.0 ``` ## Key Highlights LiteLLM v1.72.0-stable.rc is live now. Here are the key highlights of this release: - **Vector Store Permissions**: Control Vector Store access at the Key, Team, and Organization level. - **Rate Limiting Sliding Window support**: Improved accuracy for Key/Team/User rate limits with request tracking across minutes. - **Aiohttp Transport used by default**: Aiohttp transport is now the default transport for LiteLLM networking requests. This gives users 2x higher RPS per instance with a 40ms median latency overhead. - **Bedrock Agents**: Call Bedrock Agents with `/chat/completions`, `/response` endpoints. - **Anthropic File API**: Upload and analyze CSV files with Claude-4 on Anthropic via LiteLLM. - **Prometheus**: End users (`end_user`) will no longer be tracked by default on Prometheus. Tracking end_users on prometheus is now opt-in. This is done to prevent the response from `/metrics` from becoming too large. [Read More](../../docs/proxy/prometheus#tracking-end_user-on-prometheus) --- ## Vector Store Permissions This release brings support for managing permissions for vector stores by Keys, Teams, Organizations (entities) on LiteLLM. When a request attempts to query a vector store, LiteLLM will block it if the requesting entity lacks the proper permissions. This is great for use cases that require access to restricted data that you don't want everyone to use. Over the next week we plan on adding permission management for MCP Servers. --- ## Aiohttp Transport used by default Aiohttp transport is now the default transport for LiteLLM networking requests. This gives users 2x higher RPS per instance with a 40ms median latency overhead. This has been live on LiteLLM Cloud for a week + gone through alpha users testing for a week. If you encounter any issues, you can disable using the aiohttp transport in the following ways: **On LiteLLM Proxy** Set the `DISABLE_AIOHTTP_TRANSPORT=True` in the environment variables. ```yaml showLineNumbers title="Environment Variable" export DISABLE_AIOHTTP_TRANSPORT="True" ``` **On LiteLLM Python SDK** Set the `disable_aiohttp_transport=True` to disable aiohttp transport. ```python showLineNumbers title="Python SDK" import litellm litellm.disable_aiohttp_transport = True # default is False, enable this to disable aiohttp transport result = litellm.completion( model="openai/gpt-4o", messages=[{"role": "user", "content": "Hello, world!"}], ) print(result) ``` --- ## New Models / Updated Models - **[Bedrock](../../docs/providers/bedrock)** - Video support for Bedrock Converse - [PR](https://github.com/BerriAI/litellm/pull/11166) - InvokeAgents support as /chat/completions route - [PR](https://github.com/BerriAI/litellm/pull/11239), [Get Started](../../docs/providers/bedrock_agents) - AI21 Jamba models compatibility fixes - [PR](https://github.com/BerriAI/litellm/pull/11233) - Fixed duplicate maxTokens parameter for Claude with thinking - [PR](https://github.com/BerriAI/litellm/pull/11181) - **[Gemini (Google AI Studio + Vertex AI)](https://docs.litellm.ai/docs/providers/gemini)** - Parallel tool calling support with `parallel_tool_calls` parameter - [PR](https://github.com/BerriAI/litellm/pull/11125) - All Gemini models now support parallel function calling - [PR](https://github.com/BerriAI/litellm/pull/11225) - **[VertexAI](../../docs/providers/vertex)** - codeExecution tool support and anyOf handling - [PR](https://github.com/BerriAI/litellm/pull/11195) - Vertex AI Anthropic support on /v1/messages - [PR](https://github.com/BerriAI/litellm/pull/11246) - Thinking, global regions, and parallel tool calling improvements - [PR](https://github.com/BerriAI/litellm/pull/11194) - Web Search Support [PR](https://github.com/BerriAI/litellm/commit/06484f6e5a7a2f4e45c490266782ed28b51b7db6) - **[Anthropic](../../docs/providers/anthropic)** - Thinking blocks on streaming support - [PR](https://github.com/BerriAI/litellm/pull/11194) - Files API with form-data support on passthrough - [PR](https://github.com/BerriAI/litellm/pull/11256) - File ID support on /chat/completion - [PR](https://github.com/BerriAI/litellm/pull/11256) - **[xAI](../../docs/providers/xai)** - Web Search Support [PR](https://github.com/BerriAI/litellm/commit/06484f6e5a7a2f4e45c490266782ed28b51b7db6) - **[Google AI Studio](../../docs/providers/gemini)** - Web Search Support [PR](https://github.com/BerriAI/litellm/commit/06484f6e5a7a2f4e45c490266782ed28b51b7db6) - **[Mistral](../../docs/providers/mistral)** - Updated mistral-medium prices and context sizes - [PR](https://github.com/BerriAI/litellm/pull/10729) - **[Ollama](../../docs/providers/ollama)** - Tool calls parsing on streaming - [PR](https://github.com/BerriAI/litellm/pull/11171) - **[Cohere](../../docs/providers/cohere)** - Swapped Cohere and Cohere Chat provider positioning - [PR](https://github.com/BerriAI/litellm/pull/11173) - **[Nebius AI Studio](../../docs/providers/nebius)** - New provider integration - [PR](https://github.com/BerriAI/litellm/pull/11143) ## LLM API Endpoints - **[Image Edits API](../../docs/image_generation)** - Azure support for /v1/images/edits - [PR](https://github.com/BerriAI/litellm/pull/11160) - Cost tracking for image edits endpoint (OpenAI, Azure) - [PR](https://github.com/BerriAI/litellm/pull/11186) - **[Completions API](../../docs/completion/chat)** - Codestral latency overhead tracking on /v1/completions - [PR](https://github.com/BerriAI/litellm/pull/10879) - **[Audio Transcriptions API](../../docs/audio/speech)** - GPT-4o mini audio preview pricing without date - [PR](https://github.com/BerriAI/litellm/pull/11207) - Non-default params support for audio transcription - [PR](https://github.com/BerriAI/litellm/pull/11212) - **[Responses API](../../docs/response_api)** - Session management fixes for using Non-OpenAI models - [PR](https://github.com/BerriAI/litellm/pull/11254) ## Management Endpoints / UI - **Vector Stores** - Permission management for LiteLLM Keys, Teams, and Organizations - [PR](https://github.com/BerriAI/litellm/pull/11213) - UI display of vector store permissions - [PR](https://github.com/BerriAI/litellm/pull/11277) - Vector store access controls enforcement - [PR](https://github.com/BerriAI/litellm/pull/11281) - Object permissions fixes and QA improvements - [PR](https://github.com/BerriAI/litellm/pull/11291) - **Teams** - "All proxy models" display when no models selected - [PR](https://github.com/BerriAI/litellm/pull/11187) - Removed redundant teamInfo call, using existing teamsList - [PR](https://github.com/BerriAI/litellm/pull/11051) - Improved model tags display on Keys, Teams and Org pages - [PR](https://github.com/BerriAI/litellm/pull/11022) - **SSO/SCIM** - Bug fixes for showing SCIM token on UI - [PR](https://github.com/BerriAI/litellm/pull/11220) - **General UI** - Fix "UI Session Expired. Logging out" - [PR](https://github.com/BerriAI/litellm/pull/11279) - Support for forwarding /sso/key/generate to server root path URL - [PR](https://github.com/BerriAI/litellm/pull/11165) ## Logging / Guardrails Integrations #### Logging - **[Prometheus](../../docs/proxy/prometheus)** - End users will no longer be tracked by default on Prometheus. Tracking end_users on prometheus is now opt-in. [PR](https://github.com/BerriAI/litellm/pull/11192) - **[Langfuse](../../docs/proxy/logging#langfuse)** - Performance improvements: Fixed "Max langfuse clients reached" issue - [PR](https://github.com/BerriAI/litellm/pull/11285) - **[Helicone](../../docs/observability/helicone_integration)** - Base URL support - [PR](https://github.com/BerriAI/litellm/pull/11211) - **[Sentry](../../docs/proxy/logging#sentry)** - Added sentry sample rate configuration - [PR](https://github.com/BerriAI/litellm/pull/10283) #### Guardrails - **[Bedrock Guardrails](../../docs/proxy/guardrails/bedrock)** - Streaming support for bedrock post guard - [PR](https://github.com/BerriAI/litellm/pull/11247) - Auth parameter persistence fixes - [PR](https://github.com/BerriAI/litellm/pull/11270) - **[Pangea Guardrails](../../docs/proxy/guardrails/pangea)** - Added Pangea provider to Guardrails hook - [PR](https://github.com/BerriAI/litellm/pull/10775) ## Performance / Reliability Improvements - **aiohttp Transport** - Handling for aiohttp.ClientPayloadError - [PR](https://github.com/BerriAI/litellm/pull/11162) - SSL verification settings support - [PR](https://github.com/BerriAI/litellm/pull/11162) - Rollback to httpx==0.27.0 for stability - [PR](https://github.com/BerriAI/litellm/pull/11146) - **Request Limiting** - Sliding window logic for parallel request limiter v2 - [PR](https://github.com/BerriAI/litellm/pull/11283) ## Bug Fixes - **LLM API Fixes** - Added missing request_kwargs to get_available_deployment call - [PR](https://github.com/BerriAI/litellm/pull/11202) - Fixed calling Azure O-series models - [PR](https://github.com/BerriAI/litellm/pull/11212) - Support for dropping non-OpenAI params via additional_drop_params - [PR](https://github.com/BerriAI/litellm/pull/11246) - Fixed frequency_penalty to repeat_penalty parameter mapping - [PR](https://github.com/BerriAI/litellm/pull/11284) - Fix for embedding cache hits on string input - [PR](https://github.com/BerriAI/litellm/pull/11211) - **General** - OIDC provider improvements and audience bug fix - [PR](https://github.com/BerriAI/litellm/pull/10054) - Removed AzureCredentialType restriction on AZURE_CREDENTIAL - [PR](https://github.com/BerriAI/litellm/pull/11272) - Prevention of sensitive key leakage to Langfuse - [PR](https://github.com/BerriAI/litellm/pull/11165) - Fixed healthcheck test using curl when curl not in image - [PR](https://github.com/BerriAI/litellm/pull/9737) ## New Contributors * [@agajdosi](https://github.com/agajdosi) made their first contribution in [#9737](https://github.com/BerriAI/litellm/pull/9737) * [@ketangangal](https://github.com/ketangangal) made their first contribution in [#11161](https://github.com/BerriAI/litellm/pull/11161) * [@Aktsvigun](https://github.com/Aktsvigun) made their first contribution in [#11143](https://github.com/BerriAI/litellm/pull/11143) * [@ryanmeans](https://github.com/ryanmeans) made their first contribution in [#10775](https://github.com/BerriAI/litellm/pull/10775) * [@nikoizs](https://github.com/nikoizs) made their first contribution in [#10054](https://github.com/BerriAI/litellm/pull/10054) * [@Nitro963](https://github.com/Nitro963) made their first contribution in [#11202](https://github.com/BerriAI/litellm/pull/11202) * [@Jacobh2](https://github.com/Jacobh2) made their first contribution in [#11207](https://github.com/BerriAI/litellm/pull/11207) * [@regismesquita](https://github.com/regismesquita) made their first contribution in [#10729](https://github.com/BerriAI/litellm/pull/10729) * [@Vinnie-Singleton-NN](https://github.com/Vinnie-Singleton-NN) made their first contribution in [#10283](https://github.com/BerriAI/litellm/pull/10283) * [@trashhalo](https://github.com/trashhalo) made their first contribution in [#11219](https://github.com/BerriAI/litellm/pull/11219) * [@VigneshwarRajasekaran](https://github.com/VigneshwarRajasekaran) made their first contribution in [#11223](https://github.com/BerriAI/litellm/pull/11223) * [@AnilAren](https://github.com/AnilAren) made their first contribution in [#11233](https://github.com/BerriAI/litellm/pull/11233) * [@fadil4u](https://github.com/fadil4u) made their first contribution in [#11242](https://github.com/BerriAI/litellm/pull/11242) * [@whitfin](https://github.com/whitfin) made their first contribution in [#11279](https://github.com/BerriAI/litellm/pull/11279) * [@hcoona](https://github.com/hcoona) made their first contribution in [#11272](https://github.com/BerriAI/litellm/pull/11272) * [@keyute](https://github.com/keyute) made their first contribution in [#11173](https://github.com/BerriAI/litellm/pull/11173) * [@emmanuel-ferdman](https://github.com/emmanuel-ferdman) made their first contribution in [#11230](https://github.com/BerriAI/litellm/pull/11230) ## Demo Instance Here's a Demo Instance to test changes: - Instance: https://demo.litellm.ai/ - Login Credentials: - Username: admin - Password: sk-1234 ## [Git Diff](https://github.com/BerriAI/litellm/releases)