From d50210c0e354c91b0dcffd179b1bfd80527d06b3 Mon Sep 17 00:00:00 2001 From: Krrish Dholakia Date: Mon, 26 Jan 2026 18:53:35 -0800 Subject: [PATCH 1/3] docs: document new models/endpoints --- .../release_notes/v1.81.3-stable/index.md | 399 ++++++++++++++++++ 1 file changed, 399 insertions(+) create mode 100644 docs/my-website/release_notes/v1.81.3-stable/index.md diff --git a/docs/my-website/release_notes/v1.81.3-stable/index.md b/docs/my-website/release_notes/v1.81.3-stable/index.md new file mode 100644 index 00000000000..8a07e3206af --- /dev/null +++ b/docs/my-website/release_notes/v1.81.3-stable/index.md @@ -0,0 +1,399 @@ +--- +title: "v1.81.3-stable - Performance - 25% CPU Usage Reduction" +slug: "v1-81-3" +date: 2026-01-18T10:00:00 +authors: + - name: Krrish Dholakia + title: CEO, LiteLLM + url: https://www.linkedin.com/in/krish-d/ + image_url: https://pbs.twimg.com/profile_images/1298587542745358340/DZv3Oj-h_400x400.jpg + - name: Ishaan Jaff + title: CTO, LiteLLM + url: https://www.linkedin.com/in/reffajnaahsi/ + image_url: https://pbs.twimg.com/profile_images/1613813310264340481/lz54oEiB_400x400.jpg +hide_table_of_contents: false +--- + +import Image from '@theme/IdealImage'; +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + +## Deploy this version + + + + +``` showLineNumbers title="docker run litellm" +docker run \ +-e STORE_MODEL_IN_DB=True \ +-p 4000:4000 \ +docker.litellm.ai/berriai/litellm:v1.81.3.rc.2 +``` + + + + + +``` showLineNumbers title="pip install litellm" +pip install litellm==1.81.3.rc.2 +``` + + + + +--- + +## Key Highlights +[TODO] +--- + +[TODO] + +## New Models / Updated Models + +#### New Model Support + +| Provider | Model | Features | +| -------- | ----- | -------- | +| OpenAI | `gpt-5.2-codex` | Code generation | +| Azure | `azure/gpt-5.2-codex` | Code generation | +| Cerebras | `cerebras/zai-glm-4.7` | Reasoning, function calling | +| Replicate | All chat models | Full support for all Replicate chat models | + + +#### Features + +- **[OpenAI](../../docs/providers/openai)** + - Add gpt-audio and gpt-audio-mini models to pricing - [PR #19509](https://github.com/BerriAI/litellm/pull/19509) + - correct audio token costs for gpt-4o-audio-preview models - [PR #19500](https://github.com/BerriAI/litellm/pull/19500) + - Limit stop sequence as per openai spec (ensures JetBrains IDE compatibility) - [PR #19562](https://github.com/BerriAI/litellm/pull/19562) + +- **[VertexAI](../../docs/providers/vertex)** + - Docs - Google Workload Identity Federation (WIF) support - [PR #19320](https://github.com/BerriAI/litellm/pull/19320) +- **[Agentcore](../../docs/providers/bedrock_agentcore)** + - Fixes streaming issues with AWS Bedrock AgentCore where responses would stop after the first chunk, particularly affecting OAuth-enabled agents - [PR #17141](https://github.com/BerriAI/litellm/pull/17141) + +- **[Chatgpt](../../docs/providers/chatgpt)** + - Adds support for calling chatgpt subscription via LiteLLM - [PR #19030](https://github.com/BerriAI/litellm/pull/19030) + - Adds responses API bridge support for chatgpt subscription provider - [PR #19030](https://github.com/BerriAI/litellm/pull/19030) + +- **[Bedrock](../../docs/providers/bedrock)** + - support for output format for bedrock invoke via v1/messages - [PR #19560](https://github.com/BerriAI/litellm/pull/19560) + +- **[Azure](../../docs/providers/azure/azure)** + - Add support for Azure OpenAI v1 API - [PR #19313](https://github.com/BerriAI/litellm/pull/19313) + - preserve content_policy_violation details for images (#19328) - [PR #19372](https://github.com/BerriAI/litellm/pull/19372) + - Support OpenAI-format nested tool definitions for Responses API - [PR #19526](https://github.com/BerriAI/litellm/pull/19526) + +- **Gemini([Vertex AI](../../docs/providers/vertex), [Google AI Studio](../../docs/providers/gemini))** + - use responseJsonSchema for Gemini 2.0+ models - [PR #19314](https://github.com/BerriAI/litellm/pull/19314) + - Update Gemini 2.0 Flash deprecation dates to March 31, 2026 - [PR #19592](https://github.com/BerriAI/litellm/pull/19592) + +- **[Volcengine](../../docs/providers/volcano)** + - Support Volcengine responses api - [PR #18508](https://github.com/BerriAI/litellm/pull/18508) + +- **[Anthropic](../../docs/providers/anthropic)** + - Add Support for calling Claude Code Max subscriptions via LiteLLM - [PR #19453](https://github.com/BerriAI/litellm/pull/19453) + - Add Structured output for /v1/messages with Anthropic API, Azure Anthropic API, Bedrock Converse - [PR #19545](https://github.com/BerriAI/litellm/pull/19545) + +- **[Brave Search](../../docs/search/brave)** + - New Search provider - [PR #19433](https://github.com/BerriAI/litellm/pull/19433) + +- **Sarvam ai** [TODO: MISSING DOCS] + - Add support for new sarvam models - [PR #19479](https://github.com/BerriAI/litellm/pull/19479) + +- **[GMI](../../docs/providers/gmi)** + - add GMI Cloud provider support - [PR #19376](https://github.com/BerriAI/litellm/pull/19376) + + +#### Bug Fixes + +- **[Anthropic](../../docs/providers/anthropic)** + - Fix anthropic-beta sent client side being overridden instead of appended to - by @Sameerlite in #19343 + - Filter out unsupported fields from JSON schema for Anthropic's output_format API by @Sameerlite in #19482 +- **[Bedrock](../../docs/providers/bedrock)** + - Expose stability models via /image_edits endpoint and ensure proper request transformation by @Sameerlite in #19323 + - Claude Code x Bedrock Invoke fails with advanced-tool-use-2025-11-20 by @ishaan-jaff in #19373 + - deduplicate tool calls in assistant history by @binbandit in #19324 + - fix: correct us.anthropic.claude-opus-4-5 In-region pricing by @cluebbehusen in #19310 + - Fix request validation errors when using Claude 4 via bedrock invoke by @flex-myeonghyeon in #19381 + - Handle thinking with tool calls for Claude 4 models by @bensi94 in #18787 + - correct streaming choice index for tool calls by @joaodinissf in #19506 +- **[Groq](../../docs/providers/groq)** + - correct Groq gpt-oss pricing and add cache pricing by @cluebbehusen in #19311 +- **[Ollama](../../docs/providers/ollama)** + - Fix tool call errors due with improved message extraction by @Lucky-Lodhi2004 in #19369 +- **[VertexAI](../../docs/providers/vertex)** + - Removed optional vertex_count_tokens_location param before request is sent to vertex by @otaviofbrito in #19359 +- **Gemini([Vertex AI](../../docs/providers/vertex), [Google AI Studio](../../docs/providers/gemini))** + - Supports setting media_resolution and fps parameters on each video file, when using Gemini video understanding by @krisxia0506 in #19273 + - Add gemini-2.5-computer-use-preview-10-2025 model for vertex ai provider by @Sameerlite in #19397 + - handle reasoning_effort as dict from OpenAI Agents SDK by @Chesars in #19419 + - add file content support in tool results by @rynecarbone in #19416 +- **[Azure](../../docs/providers/azure_ai)** + - Fix Azure AI costs for Anthropic models by @wwwillchen in #19530 + - [Fix] Anthropic models on Azure AI cache pricing (#19532) by @jgreek in #19614 +- **[Giga Chat](../../docs/providers/gigachat)** + - Add tool choice mapping by @Sameerlite in #19645 +--- + +## LLM API Endpoints + +#### Features + +- **[/messages (Claude Code)](../../docs/providers/anthropic)** + - Add support for Tool Search on `/messages` API across Azure, Bedrock, and Anthropic API - [PR #19165](https://github.com/BerriAI/litellm/pull/19165) + - Track end-users with Claude Code (`/messages`) for better analytics and monitoring - [PR #19171](https://github.com/BerriAI/litellm/pull/19171) + - Add web search support using LiteLLM `/search` endpoint with Claude Code (`/messages`) - [PR #19263](https://github.com/BerriAI/litellm/pull/19263), [PR #19294](https://github.com/BerriAI/litellm/pull/19294) + +- **[/messages (Claude Code) - Bedrock](../../docs/providers/bedrock)** + - Add support for Prompt Caching with Bedrock Converse on `/messages` - [PR #19123](https://github.com/BerriAI/litellm/pull/19123) + - Ensure budget tokens are passed to Bedrock Converse API correctly on `/messages` - [PR #19107](https://github.com/BerriAI/litellm/pull/19107) + +- **[Responses API](../../docs/response_api)** + - Add support for caching for responses API - [PR #19068](https://github.com/BerriAI/litellm/pull/19068) + - Add retry policy support to responses API - [PR #19074](https://github.com/BerriAI/litellm/pull/19074) + +- **Realtime API** + - Use non-streaming method for endpoint v1/a2a/message/send - [PR #19025](https://github.com/BerriAI/litellm/pull/19025) + +- **Batch API** + - Fix batch deletion and retrieve - [PR #18340](https://github.com/BerriAI/litellm/pull/18340) + +#### Bugs + +- **General** + - Fix responses content can't be none - [PR #19064](https://github.com/BerriAI/litellm/pull/19064) + - Fix model name from query param in realtime request - [PR #19135](https://github.com/BerriAI/litellm/pull/19135) + - Fix video status/content credential injection for wildcard models - [PR #18854](https://github.com/BerriAI/litellm/pull/18854) + +--- + +## Management Endpoints / UI + +#### Features + +**Virtual Keys** +- View deleted keys for audit purposes - [PR #18228](https://github.com/BerriAI/litellm/pull/18228), [PR #19268](https://github.com/BerriAI/litellm/pull/19268) +- Add status query parameter for keys list - [PR #19260](https://github.com/BerriAI/litellm/pull/19260) +- Refetch keys after key creation - [PR #18994](https://github.com/BerriAI/litellm/pull/18994) +- Refresh keys list on delete - [PR #19262](https://github.com/BerriAI/litellm/pull/19262) +- Simplify key generate permission error - [PR #18997](https://github.com/BerriAI/litellm/pull/18997) +- Add search to key edit team dropdown - [PR #19119](https://github.com/BerriAI/litellm/pull/19119) + +**Teams & Organizations** +- View deleted teams for audit purposes - [PR #18228](https://github.com/BerriAI/litellm/pull/18228), [PR #19268](https://github.com/BerriAI/litellm/pull/19268) +- Add filters to organization table - [PR #18916](https://github.com/BerriAI/litellm/pull/18916) +- Add query parameters to `/organization/list` - [PR #18910](https://github.com/BerriAI/litellm/pull/18910) +- Add status query parameter for teams list - [PR #19260](https://github.com/BerriAI/litellm/pull/19260) +- Show internal users their spend only - [PR #19227](https://github.com/BerriAI/litellm/pull/19227) +- Allow preventing team admins from deleting members from teams - [PR #19128](https://github.com/BerriAI/litellm/pull/19128) +- Refactor team member icon buttons - [PR #19192](https://github.com/BerriAI/litellm/pull/19192) + +**Models + Endpoints** +- Display health information in public model hub - [PR #19256](https://github.com/BerriAI/litellm/pull/19256), [PR #19258](https://github.com/BerriAI/litellm/pull/19258) +- Quality of life improvements for Anthropic models - [PR #19058](https://github.com/BerriAI/litellm/pull/19058) +- Create reusable model select component - [PR #19164](https://github.com/BerriAI/litellm/pull/19164) +- Edit settings model dropdown - [PR #19186](https://github.com/BerriAI/litellm/pull/19186) +- Fix model hub client side exception - [PR #19045](https://github.com/BerriAI/litellm/pull/19045) + +**Usage & Analytics** +- Allow top virtual keys and models to show more entries - [PR #19050](https://github.com/BerriAI/litellm/pull/19050) +- Fix Y axis on model activity chart - [PR #19055](https://github.com/BerriAI/litellm/pull/19055) +- Add Team ID and Team Name in export report - [PR #19047](https://github.com/BerriAI/litellm/pull/19047) +- Add user metrics for Prometheus - [PR #18785](https://github.com/BerriAI/litellm/pull/18785) + +**SSO & Auth** +- Allow setting custom MSFT Base URLs - [PR #18977](https://github.com/BerriAI/litellm/pull/18977) +- Allow overriding env var attribute names - [PR #18998](https://github.com/BerriAI/litellm/pull/18998) +- Fix SCIM GET /Users error and enforce SCIM 2.0 compliance - [PR #17420](https://github.com/BerriAI/litellm/pull/17420) +- Feature flag for SCIM compliance fix - [PR #18878](https://github.com/BerriAI/litellm/pull/18878) + +**General UI** +- Add allowClear to dropdown components for better UX - [PR #18778](https://github.com/BerriAI/litellm/pull/18778) +- Add community engagement buttons - [PR #19114](https://github.com/BerriAI/litellm/pull/19114) +- UI Feedback Form - why LiteLLM - [PR #18999](https://github.com/BerriAI/litellm/pull/18999) +- Refactor user and team table filters to reusable component - [PR #19010](https://github.com/BerriAI/litellm/pull/19010) +- Adjusting new badges - [PR #19278](https://github.com/BerriAI/litellm/pull/19278) + +#### Bugs + +- Container API routes return 401 for non-admin users - routes missing from openai_routes - [PR #19115](https://github.com/BerriAI/litellm/pull/19115) +- Allow routing to regional endpoints for Containers API - [PR #19118](https://github.com/BerriAI/litellm/pull/19118) +- Fix Azure Storage circular reference error - [PR #19120](https://github.com/BerriAI/litellm/pull/19120) +- Fix prompt deletion fails with Prisma FieldNotFoundError - [PR #18966](https://github.com/BerriAI/litellm/pull/18966) + +--- + +## AI Integrations + +### Logging + +- **[OpenTelemetry](../../docs/proxy/logging#opentelemetry)** + - Update semantic conventions to 1.38 (gen_ai attributes) - [PR #18793](https://github.com/BerriAI/litellm/pull/18793) + +- **[LangSmith](../../docs/proxy/logging#langsmith)** + - Hoist thread grouping metadata (session_id, thread) - [PR #18982](https://github.com/BerriAI/litellm/pull/18982) + +- **[Langfuse](../../docs/proxy/logging#langfuse)** + - Include Langfuse logger in JSON logging when Langfuse callback is used - [PR #19162](https://github.com/BerriAI/litellm/pull/19162) + +- **[Logfire](../../docs/observability/logfire)** + - Add ability to customize Logfire base URL through env var - [PR #19148](https://github.com/BerriAI/litellm/pull/19148) + +- **General Logging** + - Enable JSON logging via configuration and add regression test - [PR #19037](https://github.com/BerriAI/litellm/pull/19037) + - Fix header forwarding for embeddings endpoint - [PR #18960](https://github.com/BerriAI/litellm/pull/18960) + - Preserve llm_provider-* headers in error responses - [PR #19020](https://github.com/BerriAI/litellm/pull/19020) + - Fix turn_off_message_logging not redacting request messages in proxy_server_request field - [PR #18897](https://github.com/BerriAI/litellm/pull/18897) + +### Guardrails + +- **[Grayswan](../../docs/proxy/guardrails/grayswan)** + - Implement fail-open option (default: True) - [PR #18266](https://github.com/BerriAI/litellm/pull/18266) + +- **[Pangea](../../docs/proxy/guardrails/pangea)** + - Respect `default_on` during initialization - [PR #18912](https://github.com/BerriAI/litellm/pull/18912) + +- **[Panw Prisma AIRS](../../docs/proxy/guardrails/panw_prisma_airs)** + - Add custom violation message support - [PR #19272](https://github.com/BerriAI/litellm/pull/19272) + +- **General Guardrails** + - Fix SerializationIterator error and pass tools to guardrail - [PR #18932](https://github.com/BerriAI/litellm/pull/18932) + - Properly handle custom guardrails parameters - [PR #18978](https://github.com/BerriAI/litellm/pull/18978) + - Use clean error messages for blocked requests - [PR #19023](https://github.com/BerriAI/litellm/pull/19023) + - Guardrail moderation support with responses API - [PR #18957](https://github.com/BerriAI/litellm/pull/18957) + - Fix model-level guardrails not taking effect - [PR #18895](https://github.com/BerriAI/litellm/pull/18895) + +--- + +## Spend Tracking, Budgets and Rate Limiting + +- **Cost Calculation Fixes** + - Include IMAGE token count in cost calculation for Gemini models - [PR #18876](https://github.com/BerriAI/litellm/pull/18876) + - Fix negative text_tokens when using cache with images - [PR #18768](https://github.com/BerriAI/litellm/pull/18768) + - Fix image tokens spend logging for `/images/generations` - [PR #19009](https://github.com/BerriAI/litellm/pull/19009) + - Fix incorrect `prompt_tokens_details` in Gemini Image Generation - [PR #19070](https://github.com/BerriAI/litellm/pull/19070) + - Fix case-insensitive model cost map lookup - [PR #18208](https://github.com/BerriAI/litellm/pull/18208) + +- **Pricing Updates** + - Correct pricing for `openrouter/openai/gpt-oss-20b` - [PR #18899](https://github.com/BerriAI/litellm/pull/18899) + - Add pricing for `azure_ai/claude-opus-4-5` - [PR #19003](https://github.com/BerriAI/litellm/pull/19003) + - Update Novita models prices - [PR #19005](https://github.com/BerriAI/litellm/pull/19005) + - Fix Azure Grok prices - [PR #19102](https://github.com/BerriAI/litellm/pull/19102) + - Fix GCP GLM-4.7 pricing - [PR #19172](https://github.com/BerriAI/litellm/pull/19172) + - Sync DeepSeek chat/reasoner to V3.2 pricing - [PR #18884](https://github.com/BerriAI/litellm/pull/18884) + - Correct cache_read pricing for gemini-2.5-pro models - [PR #18157](https://github.com/BerriAI/litellm/pull/18157) + +- **Budget & Rate Limiting** + - Correct budget limit validation operator (>=) for team members - [PR #19207](https://github.com/BerriAI/litellm/pull/19207) + - Fix TPM 25% limiting by ensuring priority queue logic - [PR #19092](https://github.com/BerriAI/litellm/pull/19092) + - Cleanup spend logs cron verification, fix, and docs - [PR #19085](https://github.com/BerriAI/litellm/pull/19085) + +--- + +## MCP Gateway + +- Prevent duplicate MCP reload scheduler registration - [PR #18934](https://github.com/BerriAI/litellm/pull/18934) +- Forward MCP extra headers case-insensitively - [PR #18940](https://github.com/BerriAI/litellm/pull/18940) +- Fix MCP REST auth checks - [PR #19051](https://github.com/BerriAI/litellm/pull/19051) +- Fix generating two telemetry events in responses - [PR #18938](https://github.com/BerriAI/litellm/pull/18938) +- Fix MCP chat completions - [PR #19129](https://github.com/BerriAI/litellm/pull/19129) + +--- + +## Performance / Loadbalancing / Reliability improvements + +- **Performance Improvements** + - Remove bottleneck causing high CPU usage & overhead under heavy load - [PR #19049](https://github.com/BerriAI/litellm/pull/19049) + - Add CI enforcement for O(1) operations in `_get_model_cost_key` to prevent performance regressions - [PR #19052](https://github.com/BerriAI/litellm/pull/19052) + - Fix Azure embeddings JSON parsing to prevent connection leaks and ensure proper router cooldown - [PR #19167](https://github.com/BerriAI/litellm/pull/19167) + - Do not fallback to token counter if `disable_token_counter` is enabled - [PR #19041](https://github.com/BerriAI/litellm/pull/19041) + +- **Reliability** + - Add fallback endpoints support - [PR #19185](https://github.com/BerriAI/litellm/pull/19185) + - Fix stream_timeout parameter functionality - [PR #19191](https://github.com/BerriAI/litellm/pull/19191) + - Fix model matching priority in configuration - [PR #19012](https://github.com/BerriAI/litellm/pull/19012) + - Fix num_retries in litellm_params as per config - [PR #18975](https://github.com/BerriAI/litellm/pull/18975) + - Handle exceptions without response parameter - [PR #18919](https://github.com/BerriAI/litellm/pull/18919) + +- **Infrastructure** + - Add Custom CA certificates to boto3 clients - [PR #18942](https://github.com/BerriAI/litellm/pull/18942) + - Update boto3 to 1.40.15 and aioboto3 to 15.5.0 - [PR #19090](https://github.com/BerriAI/litellm/pull/19090) + - Make keepalive_timeout parameter work for Gunicorn - [PR #19087](https://github.com/BerriAI/litellm/pull/19087) + +- **Helm Chart** + - Fix mount config.yaml as single file in Helm chart - [PR #19146](https://github.com/BerriAI/litellm/pull/19146) + - Sync Helm chart versioning with production standards and Docker versions - [PR #18868](https://github.com/BerriAI/litellm/pull/18868) + +--- + +## Database Changes + +### Schema Updates + +| Table | Change Type | Description | PR | +| ----- | ----------- | ----------- | -- | +| `LiteLLM_ProxyModelTable` | New Columns | Added `created_at` and `updated_at` timestamp fields | [PR #18937](https://github.com/BerriAI/litellm/pull/18937) | + +--- + +## Documentation Updates + +- Add LiteLLM architecture md doc - [PR #19057](https://github.com/BerriAI/litellm/pull/19057), [PR #19252](https://github.com/BerriAI/litellm/pull/19252) +- Add troubleshooting guide - [PR #19096](https://github.com/BerriAI/litellm/pull/19096), [PR #19097](https://github.com/BerriAI/litellm/pull/19097), [PR #19099](https://github.com/BerriAI/litellm/pull/19099) +- Add structured issue reporting guides for CPU and memory issues - [PR #19117](https://github.com/BerriAI/litellm/pull/19117) +- Add Redis requirement warning for high-traffic deployments - [PR #18892](https://github.com/BerriAI/litellm/pull/18892) +- Update load balancing and routing with enable_pre_call_checks - [PR #18888](https://github.com/BerriAI/litellm/pull/18888) +- Updated pass_through with guided param - [PR #18886](https://github.com/BerriAI/litellm/pull/18886) +- Update message content types link and add content types table - [PR #18209](https://github.com/BerriAI/litellm/pull/18209) +- Add Redis initialization with kwargs - [PR #19183](https://github.com/BerriAI/litellm/pull/19183) +- Improve documentation for routing LLM calls via SAP Gen AI Hub - [PR #19166](https://github.com/BerriAI/litellm/pull/19166) +- Deleted Keys and Teams docs - [PR #19291](https://github.com/BerriAI/litellm/pull/19291) +- Claude Code end user tracking guide - [PR #19176](https://github.com/BerriAI/litellm/pull/19176) +- Add MCP troubleshooting guide - [PR #19122](https://github.com/BerriAI/litellm/pull/19122) +- Add auth message UI documentation - [PR #19063](https://github.com/BerriAI/litellm/pull/19063) +- Add guide for mounting custom callbacks in Helm/K8s - [PR #19136](https://github.com/BerriAI/litellm/pull/19136) + +--- + +## Bug Fixes + +- Fix Swagger UI path execute error with server_root_path in OpenAPI schema - [PR #18947](https://github.com/BerriAI/litellm/pull/18947) +- Normalize OpenAI SDK BaseModel choices/messages to avoid Pydantic serializer warnings - [PR #18972](https://github.com/BerriAI/litellm/pull/18972) +- Add contextual gap checks and word-form digits - [PR #18301](https://github.com/BerriAI/litellm/pull/18301) +- Clean up orphaned files from repository root - [PR #19150](https://github.com/BerriAI/litellm/pull/19150) +- Include proxy/prisma_migration.py in non-root - [PR #18971](https://github.com/BerriAI/litellm/pull/18971) +- Update prisma_migration.py - [PR #19083](https://github.com/BerriAI/litellm/pull/19083) + +--- + +## New Contributors + +* @yogeshwaran10 made their first contribution in [PR #18898](https://github.com/BerriAI/litellm/pull/18898) +* @theonlypal made their first contribution in [PR #18937](https://github.com/BerriAI/litellm/pull/18937) +* @jonmagic made their first contribution in [PR #18935](https://github.com/BerriAI/litellm/pull/18935) +* @houdataali made their first contribution in [PR #19025](https://github.com/BerriAI/litellm/pull/19025) +* @hummat made their first contribution in [PR #18972](https://github.com/BerriAI/litellm/pull/18972) +* @berkeyalciin made their first contribution in [PR #18966](https://github.com/BerriAI/litellm/pull/18966) +* @MateuszOssGit made their first contribution in [PR #18959](https://github.com/BerriAI/litellm/pull/18959) +* @xfan001 made their first contribution in [PR #18947](https://github.com/BerriAI/litellm/pull/18947) +* @nulone made their first contribution in [PR #18884](https://github.com/BerriAI/litellm/pull/18884) +* @debnil-mercor made their first contribution in [PR #18919](https://github.com/BerriAI/litellm/pull/18919) +* @hakhundov made their first contribution in [PR #17420](https://github.com/BerriAI/litellm/pull/17420) +* @rohanwinsor made their first contribution in [PR #19078](https://github.com/BerriAI/litellm/pull/19078) +* @pgolm made their first contribution in [PR #19020](https://github.com/BerriAI/litellm/pull/19020) +* @vikigenius made their first contribution in [PR #19148](https://github.com/BerriAI/litellm/pull/19148) +* @burnerburnerburnerman made their first contribution in [PR #19090](https://github.com/BerriAI/litellm/pull/19090) +* @yfge made their first contribution in [PR #19076](https://github.com/BerriAI/litellm/pull/19076) +* @danielnyari-seon made their first contribution in [PR #19083](https://github.com/BerriAI/litellm/pull/19083) +* @guilherme-segantini made their first contribution in [PR #19166](https://github.com/BerriAI/litellm/pull/19166) +* @jgreek made their first contribution in [PR #19147](https://github.com/BerriAI/litellm/pull/19147) +* @anand-kamble made their first contribution in [PR #19193](https://github.com/BerriAI/litellm/pull/19193) +* @neubig made their first contribution in [PR #19162](https://github.com/BerriAI/litellm/pull/19162) + +--- + +## Full Changelog + +**[View complete changelog on GitHub](https://github.com/BerriAI/litellm/compare/v1.80.15.rc.1...v1.81.0.rc.1)** From 25a558bb7320cbfadb341d3365426ccc787f7f88 Mon Sep 17 00:00:00 2001 From: Krrish Dholakia Date: Mon, 26 Jan 2026 19:31:14 -0800 Subject: [PATCH 2/3] docs: cleanup --- .../release_notes/v1.81.3-stable/index.md | 465 +++++++++--------- 1 file changed, 244 insertions(+), 221 deletions(-) diff --git a/docs/my-website/release_notes/v1.81.3-stable/index.md b/docs/my-website/release_notes/v1.81.3-stable/index.md index 8a07e3206af..6c3f945b9e6 100644 --- a/docs/my-website/release_notes/v1.81.3-stable/index.md +++ b/docs/my-website/release_notes/v1.81.3-stable/index.md @@ -1,7 +1,7 @@ --- title: "v1.81.3-stable - Performance - 25% CPU Usage Reduction" slug: "v1-81-3" -date: 2026-01-18T10:00:00 +date: 2026-01-26T10:00:00 authors: - name: Krrish Dholakia title: CEO, LiteLLM @@ -51,7 +51,7 @@ pip install litellm==1.81.3.rc.2 ## New Models / Updated Models -#### New Model Support +### New Model Support | Provider | Model | Features | | -------- | ----- | -------- | @@ -61,7 +61,7 @@ pip install litellm==1.81.3.rc.2 | Replicate | All chat models | Full support for all Replicate chat models | -#### Features +### Features - **[OpenAI](../../docs/providers/openai)** - Add gpt-audio and gpt-audio-mini models to pricing - [PR #19509](https://github.com/BerriAI/litellm/pull/19509) @@ -106,294 +106,317 @@ pip install litellm==1.81.3.rc.2 - add GMI Cloud provider support - [PR #19376](https://github.com/BerriAI/litellm/pull/19376) -#### Bug Fixes +### Bug Fixes - **[Anthropic](../../docs/providers/anthropic)** - - Fix anthropic-beta sent client side being overridden instead of appended to - by @Sameerlite in #19343 - - Filter out unsupported fields from JSON schema for Anthropic's output_format API by @Sameerlite in #19482 + - Fix anthropic-beta sent client side being overridden instead of appended to - [PR #19343](https://github.com/BerriAI/litellm/pull/19343) + - Filter out unsupported fields from JSON schema for Anthropic's output_format API - [PR #19482](https://github.com/BerriAI/litellm/pull/19482) - **[Bedrock](../../docs/providers/bedrock)** - - Expose stability models via /image_edits endpoint and ensure proper request transformation by @Sameerlite in #19323 - - Claude Code x Bedrock Invoke fails with advanced-tool-use-2025-11-20 by @ishaan-jaff in #19373 - - deduplicate tool calls in assistant history by @binbandit in #19324 - - fix: correct us.anthropic.claude-opus-4-5 In-region pricing by @cluebbehusen in #19310 - - Fix request validation errors when using Claude 4 via bedrock invoke by @flex-myeonghyeon in #19381 - - Handle thinking with tool calls for Claude 4 models by @bensi94 in #18787 - - correct streaming choice index for tool calls by @joaodinissf in #19506 + - Expose stability models via /image_edits endpoint and ensure proper request transformation - [PR #19323](https://github.com/BerriAI/litellm/pull/19323) + - Claude Code x Bedrock Invoke fails with advanced-tool-use-2025-11-20 - [PR #19373](https://github.com/BerriAI/litellm/pull/19373) + - deduplicate tool calls in assistant history - [PR #19324](https://github.com/BerriAI/litellm/pull/19324) + - fix: correct us.anthropic.claude-opus-4-5 In-region pricing - [PR #19310](https://github.com/BerriAI/litellm/pull/19310) + - Fix request validation errors when using Claude 4 via bedrock invoke - [PR #19381](https://github.com/BerriAI/litellm/pull/19381) + - Handle thinking with tool calls for Claude 4 models - [PR #19506](https://github.com/BerriAI/litellm/pull/19506) + - correct streaming choice index for tool calls - [PR #19506](https://github.com/BerriAI/litellm/pull/19506) - **[Groq](../../docs/providers/groq)** - - correct Groq gpt-oss pricing and add cache pricing by @cluebbehusen in #19311 + - correct Groq gpt-oss pricing and add cache pricing - [PR #19311](https://github.com/BerriAI/litellm/pull/19311) - **[Ollama](../../docs/providers/ollama)** - - Fix tool call errors due with improved message extraction by @Lucky-Lodhi2004 in #19369 + - Fix tool call errors due with improved message extraction - [PR #19369](https://github.com/BerriAI/litellm/pull/19369) - **[VertexAI](../../docs/providers/vertex)** - - Removed optional vertex_count_tokens_location param before request is sent to vertex by @otaviofbrito in #19359 + - Removed optional vertex_count_tokens_location param before request is sent to vertex - [PR #19359](https://github.com/BerriAI/litellm/pull/19359) - **Gemini([Vertex AI](../../docs/providers/vertex), [Google AI Studio](../../docs/providers/gemini))** - - Supports setting media_resolution and fps parameters on each video file, when using Gemini video understanding by @krisxia0506 in #19273 - - Add gemini-2.5-computer-use-preview-10-2025 model for vertex ai provider by @Sameerlite in #19397 - - handle reasoning_effort as dict from OpenAI Agents SDK by @Chesars in #19419 - - add file content support in tool results by @rynecarbone in #19416 + - Supports setting media_resolution and fps parameters on each video file, when using Gemini video understanding - [PR #19273](https://github.com/BerriAI/litellm/pull/19273) + - Add gemini-2.5-computer-use-preview-10-2025 model for vertex ai provider - [PR #19397](https://github.com/BerriAI/litellm/pull/19397) + - handle reasoning_effort as dict from OpenAI Agents SDK - [PR #19419](https://github.com/BerriAI/litellm/pull/19419) + - add file content support in tool results - [PR #19416](https://github.com/BerriAI/litellm/pull/19416) - **[Azure](../../docs/providers/azure_ai)** - - Fix Azure AI costs for Anthropic models by @wwwillchen in #19530 - - [Fix] Anthropic models on Azure AI cache pricing (#19532) by @jgreek in #19614 + - Fix Azure AI costs for Anthropic models - [PR #19530](https://github.com/BerriAI/litellm/pull/19530) + - Anthropic models on Azure AI cache pricing - [PR #19532](https://github.com/BerriAI/litellm/pull/19532) - **[Giga Chat](../../docs/providers/gigachat)** - - Add tool choice mapping by @Sameerlite in #19645 + - Add tool choice mapping - [PR #19645](https://github.com/BerriAI/litellm/pull/19645) --- -## LLM API Endpoints +## AI API Endpoints (LLMs, MCP, Agents) -#### Features +### Features -- **[/messages (Claude Code)](../../docs/providers/anthropic)** - - Add support for Tool Search on `/messages` API across Azure, Bedrock, and Anthropic API - [PR #19165](https://github.com/BerriAI/litellm/pull/19165) - - Track end-users with Claude Code (`/messages`) for better analytics and monitoring - [PR #19171](https://github.com/BerriAI/litellm/pull/19171) - - Add web search support using LiteLLM `/search` endpoint with Claude Code (`/messages`) - [PR #19263](https://github.com/BerriAI/litellm/pull/19263), [PR #19294](https://github.com/BerriAI/litellm/pull/19294) +- **[Files API](../../docs/files_endpoints)** + - Add managed files support when load_balancing is True - [PR #19338](https://github.com/BerriAI/litellm/pull/19338) -- **[/messages (Claude Code) - Bedrock](../../docs/providers/bedrock)** - - Add support for Prompt Caching with Bedrock Converse on `/messages` - [PR #19123](https://github.com/BerriAI/litellm/pull/19123) - - Ensure budget tokens are passed to Bedrock Converse API correctly on `/messages` - [PR #19107](https://github.com/BerriAI/litellm/pull/19107) +- **[Claude Plugin Marketplace](../../docs/tutorials/claude_code_plugin_marketplace)** + - Add self hosted Claude Code Plugin Marketplace - [PR #19378](https://github.com/BerriAI/litellm/pull/19378) + +- **[MCP](../../docs/mcp)** + - Add MCP Protocol version 2025-11-25 support - [PR #19379](https://github.com/BerriAI/litellm/pull/19379) + - Log MCP tool calls and list tools in the LiteLLM Spend Logs table for easier debugging - [PR #19469](https://github.com/BerriAI/litellm/pull/19469) + +- **[Vertex AI](../../docs/providers/vertex)** + - Ensure only anthropic betas are forwarded down to LLM API (by default) - [PR #19542](https://github.com/BerriAI/litellm/pull/19542) + - Allow overriding to support forwarding incoming headers are forwarded down to target - [PR #19524](https://github.com/BerriAI/litellm/pull/19524) + +- **[Chat/Completions](../../docs/completion/input)** + - Add MCP tools response to chat completions - [PR #19552](https://github.com/BerriAI/litellm/pull/19552) + - Add custom vertex ai finish reasons to the output - [PR #19558](https://github.com/BerriAI/litellm/pull/19558) + - Return MCP execution in /chat/completions before model output during streaming - [PR #19623](https://github.com/BerriAI/litellm/pull/19623) + +### Bugs - **[Responses API](../../docs/response_api)** - - Add support for caching for responses API - [PR #19068](https://github.com/BerriAI/litellm/pull/19068) - - Add retry policy support to responses API - [PR #19074](https://github.com/BerriAI/litellm/pull/19074) + - Fix duplicate messages during MCP streaming tool execution - [PR #19317](https://github.com/BerriAI/litellm/pull/19317) + - Fix pickle error when using OpenAI's Responses API with stream=True and tool_choice of type allowed_tools (an OpenAI-native parameter) - [PR #17205](https://github.com/BerriAI/litellm/pull/17205) + - stream tool call events for non-openai models - [PR #19368](https://github.com/BerriAI/litellm/pull/19368) + - preserve tool output ordering for gemini in responses bridge - [PR #19360](https://github.com/BerriAI/litellm/pull/19360) + - Add ID caching to prevent ID mismatch text-start and text-delta - [PR #19390](https://github.com/BerriAI/litellm/pull/19390) + - Include output_item, reasoning_summary_Text_done and reasoning_summary_part_done events for non-openai models - [PR #19472](https://github.com/BerriAI/litellm/pull/19472) -- **Realtime API** - - Use non-streaming method for endpoint v1/a2a/message/send - [PR #19025](https://github.com/BerriAI/litellm/pull/19025) +- **[Chat/Completions](../../docs/completion/input)** + - fix: drop_params not dropping prompt_cache_key for non-OpenAI providers - [PR #19346](https://github.com/BerriAI/litellm/pull/19346) -- **Batch API** - - Fix batch deletion and retrieve - [PR #18340](https://github.com/BerriAI/litellm/pull/18340) +- **[Realtime API](../../docs/realtime)** + - disable SSL for ws:// WebSocket connections - [PR #19345](https://github.com/BerriAI/litellm/pull/19345) -#### Bugs +- **[Generate Content](../../docs/generateContent)** + - Log actual user input when google genai/vertex endpoints are called client-side - [PR #19156](https://github.com/BerriAI/litellm/pull/19156) -- **General** - - Fix responses content can't be none - [PR #19064](https://github.com/BerriAI/litellm/pull/19064) - - Fix model name from query param in realtime request - [PR #19135](https://github.com/BerriAI/litellm/pull/19135) - - Fix video status/content credential injection for wildcard models - [PR #18854](https://github.com/BerriAI/litellm/pull/18854) +- **[/messages/count_tokens Anthropic Token Counting](../../docs/anthropic_count_tokens)** + - ensure it works for Anthropic, Azure AI Anthropic on AI Gateway - [PR #19432](https://github.com/BerriAI/litellm/pull/19432) +- **[MCP](../../docs/mcp)** + - forward static_headers to MCP servers - [PR #19366](https://github.com/BerriAI/litellm/pull/19366) + +- **[Batch API](../../docs/batches)** + - Fix: generation config empty for batch - [PR #19556](https://github.com/BerriAI/litellm/pull/19556) + +- **[Pass Through Endpoints](../../docs/proxy/pass_through)** + - Always reupdate registry - [PR #19420](https://github.com/BerriAI/litellm/pull/19420) --- ## Management Endpoints / UI -#### Features - -**Virtual Keys** -- View deleted keys for audit purposes - [PR #18228](https://github.com/BerriAI/litellm/pull/18228), [PR #19268](https://github.com/BerriAI/litellm/pull/19268) -- Add status query parameter for keys list - [PR #19260](https://github.com/BerriAI/litellm/pull/19260) -- Refetch keys after key creation - [PR #18994](https://github.com/BerriAI/litellm/pull/18994) -- Refresh keys list on delete - [PR #19262](https://github.com/BerriAI/litellm/pull/19262) -- Simplify key generate permission error - [PR #18997](https://github.com/BerriAI/litellm/pull/18997) -- Add search to key edit team dropdown - [PR #19119](https://github.com/BerriAI/litellm/pull/19119) - -**Teams & Organizations** -- View deleted teams for audit purposes - [PR #18228](https://github.com/BerriAI/litellm/pull/18228), [PR #19268](https://github.com/BerriAI/litellm/pull/19268) -- Add filters to organization table - [PR #18916](https://github.com/BerriAI/litellm/pull/18916) -- Add query parameters to `/organization/list` - [PR #18910](https://github.com/BerriAI/litellm/pull/18910) -- Add status query parameter for teams list - [PR #19260](https://github.com/BerriAI/litellm/pull/19260) -- Show internal users their spend only - [PR #19227](https://github.com/BerriAI/litellm/pull/19227) -- Allow preventing team admins from deleting members from teams - [PR #19128](https://github.com/BerriAI/litellm/pull/19128) -- Refactor team member icon buttons - [PR #19192](https://github.com/BerriAI/litellm/pull/19192) - -**Models + Endpoints** -- Display health information in public model hub - [PR #19256](https://github.com/BerriAI/litellm/pull/19256), [PR #19258](https://github.com/BerriAI/litellm/pull/19258) -- Quality of life improvements for Anthropic models - [PR #19058](https://github.com/BerriAI/litellm/pull/19058) -- Create reusable model select component - [PR #19164](https://github.com/BerriAI/litellm/pull/19164) -- Edit settings model dropdown - [PR #19186](https://github.com/BerriAI/litellm/pull/19186) -- Fix model hub client side exception - [PR #19045](https://github.com/BerriAI/litellm/pull/19045) - -**Usage & Analytics** -- Allow top virtual keys and models to show more entries - [PR #19050](https://github.com/BerriAI/litellm/pull/19050) -- Fix Y axis on model activity chart - [PR #19055](https://github.com/BerriAI/litellm/pull/19055) -- Add Team ID and Team Name in export report - [PR #19047](https://github.com/BerriAI/litellm/pull/19047) -- Add user metrics for Prometheus - [PR #18785](https://github.com/BerriAI/litellm/pull/18785) - -**SSO & Auth** -- Allow setting custom MSFT Base URLs - [PR #18977](https://github.com/BerriAI/litellm/pull/18977) -- Allow overriding env var attribute names - [PR #18998](https://github.com/BerriAI/litellm/pull/18998) -- Fix SCIM GET /Users error and enforce SCIM 2.0 compliance - [PR #17420](https://github.com/BerriAI/litellm/pull/17420) -- Feature flag for SCIM compliance fix - [PR #18878](https://github.com/BerriAI/litellm/pull/18878) - -**General UI** -- Add allowClear to dropdown components for better UX - [PR #18778](https://github.com/BerriAI/litellm/pull/18778) -- Add community engagement buttons - [PR #19114](https://github.com/BerriAI/litellm/pull/19114) -- UI Feedback Form - why LiteLLM - [PR #18999](https://github.com/BerriAI/litellm/pull/18999) -- Refactor user and team table filters to reusable component - [PR #19010](https://github.com/BerriAI/litellm/pull/19010) -- Adjusting new badges - [PR #19278](https://github.com/BerriAI/litellm/pull/19278) - -#### Bugs - -- Container API routes return 401 for non-admin users - routes missing from openai_routes - [PR #19115](https://github.com/BerriAI/litellm/pull/19115) -- Allow routing to regional endpoints for Containers API - [PR #19118](https://github.com/BerriAI/litellm/pull/19118) -- Fix Azure Storage circular reference error - [PR #19120](https://github.com/BerriAI/litellm/pull/19120) -- Fix prompt deletion fails with Prisma FieldNotFoundError - [PR #18966](https://github.com/BerriAI/litellm/pull/18966) +### Features ---- +- **Cost Estimator** + - Fix model dropdown - [PR #19529](https://github.com/BerriAI/litellm/pull/19529) -## AI Integrations +- **Claude Code Plugins** + - Allow Adding Claude Code Plugins via UI - [PR #19387](https://github.com/BerriAI/litellm/pull/19387) -### Logging +- **Guardrails** + - New Policy management UI - [PR #19668](https://github.com/BerriAI/litellm/pull/19668) + - Allow adding policies on Keys/Teams + Viewing on Info panels - [PR #19688](https://github.com/BerriAI/litellm/pull/19688) -- **[OpenTelemetry](../../docs/proxy/logging#opentelemetry)** - - Update semantic conventions to 1.38 (gen_ai attributes) - [PR #18793](https://github.com/BerriAI/litellm/pull/18793) +- **General** + - respects custom authentication header override - [PR #19276](https://github.com/BerriAI/litellm/pull/19276) -- **[LangSmith](../../docs/proxy/logging#langsmith)** - - Hoist thread grouping metadata (session_id, thread) - [PR #18982](https://github.com/BerriAI/litellm/pull/18982) +- **Playground** + - Button to Fill Custom API Base - [PR #19440](https://github.com/BerriAI/litellm/pull/19440) + - display mcp output on the play ground - [PR #19553](https://github.com/BerriAI/litellm/pull/19553) -- **[Langfuse](../../docs/proxy/logging#langfuse)** - - Include Langfuse logger in JSON logging when Langfuse callback is used - [PR #19162](https://github.com/BerriAI/litellm/pull/19162) +- **Models** + - Paginate /v2/models/info - [PR #19521](https://github.com/BerriAI/litellm/pull/19521) + - All Model Tab Pagination - [PR #19525](https://github.com/BerriAI/litellm/pull/19525) + - Adding Optional scope Param to /models - [PR #19539](https://github.com/BerriAI/litellm/pull/19539) + - Model Search - [PR #19622](https://github.com/BerriAI/litellm/pull/19622) + - Filter by Model ID and Team ID - [PR #19713](https://github.com/BerriAI/litellm/pull/19713) -- **[Logfire](../../docs/observability/logfire)** - - Add ability to customize Logfire base URL through env var - [PR #19148](https://github.com/BerriAI/litellm/pull/19148) +- **MCP Servers** + - MCP Tools Tab Resetting to Overview - [PR #19468](https://github.com/BerriAI/litellm/pull/19468) -- **General Logging** - - Enable JSON logging via configuration and add regression test - [PR #19037](https://github.com/BerriAI/litellm/pull/19037) - - Fix header forwarding for embeddings endpoint - [PR #18960](https://github.com/BerriAI/litellm/pull/18960) - - Preserve llm_provider-* headers in error responses - [PR #19020](https://github.com/BerriAI/litellm/pull/19020) - - Fix turn_off_message_logging not redacting request messages in proxy_server_request field - [PR #18897](https://github.com/BerriAI/litellm/pull/18897) +- **Organizations** + - Prevent org admin from creating a new user with proxy_admin permissions - [PR #19296](https://github.com/BerriAI/litellm/pull/19296) + - Edit Page: Reusable Model Select - [PR #19601](https://github.com/BerriAI/litellm/pull/19601) -### Guardrails +- **Teams** + - Reusable Model Select - [PR #19543](https://github.com/BerriAI/litellm/pull/19543) + - [Fix] Team Update with Organization having All Proxy Models - [PR #19604](https://github.com/BerriAI/litellm/pull/19604) + +- **Logs** + - Include tool arguments in spend logs table - [PR #19640](https://github.com/BerriAI/litellm/pull/19640) -- **[Grayswan](../../docs/proxy/guardrails/grayswan)** - - Implement fail-open option (default: True) - [PR #18266](https://github.com/BerriAI/litellm/pull/18266) +- **Fallbacks / Loadbalancing** + - New fallbacks modal - [PR #19673](https://github.com/BerriAI/litellm/pull/19673) + - Set fallbacks/loadbalancing by team/key - [PR #19686](https://github.com/BerriAI/litellm/pull/19686) -- **[Pangea](../../docs/proxy/guardrails/pangea)** - - Respect `default_on` during initialization - [PR #18912](https://github.com/BerriAI/litellm/pull/18912) +### Bugs -- **[Panw Prisma AIRS](../../docs/proxy/guardrails/panw_prisma_airs)** - - Add custom violation message support - [PR #19272](https://github.com/BerriAI/litellm/pull/19272) +- **Playground** + - increase model selector width in playground Compare view - [PR #19423](https://github.com/BerriAI/litellm/pull/19423) -- **General Guardrails** - - Fix SerializationIterator error and pass tools to guardrail - [PR #18932](https://github.com/BerriAI/litellm/pull/18932) - - Properly handle custom guardrails parameters - [PR #18978](https://github.com/BerriAI/litellm/pull/18978) - - Use clean error messages for blocked requests - [PR #19023](https://github.com/BerriAI/litellm/pull/19023) - - Guardrail moderation support with responses API - [PR #18957](https://github.com/BerriAI/litellm/pull/18957) - - Fix model-level guardrails not taking effect - [PR #18895](https://github.com/BerriAI/litellm/pull/18895) +- **Virtual Keys** + - Sorting Shows Incorrect Entries - [PR #19534](https://github.com/BerriAI/litellm/pull/19534) +- **General** + - UI 404 error when SERVER_ROOT_PATH is set - [PR #19467](https://github.com/BerriAI/litellm/pull/19467) + - Redirect to ui/login on expired JWT - [PR #19687](https://github.com/BerriAI/litellm/pull/19687) + +- **SSO** + - Fix SSO user roles not updating for existing users - [PR #19621](https://github.com/BerriAI/litellm/pull/19621) + +- **Guardrails** + - ensure guardrail patterns persist on edit and mode toggle - [PR #19265](https://github.com/BerriAI/litellm/pull/19265) --- -## Spend Tracking, Budgets and Rate Limiting +## AI Integrations -- **Cost Calculation Fixes** - - Include IMAGE token count in cost calculation for Gemini models - [PR #18876](https://github.com/BerriAI/litellm/pull/18876) - - Fix negative text_tokens when using cache with images - [PR #18768](https://github.com/BerriAI/litellm/pull/18768) - - Fix image tokens spend logging for `/images/generations` - [PR #19009](https://github.com/BerriAI/litellm/pull/19009) - - Fix incorrect `prompt_tokens_details` in Gemini Image Generation - [PR #19070](https://github.com/BerriAI/litellm/pull/19070) - - Fix case-insensitive model cost map lookup - [PR #18208](https://github.com/BerriAI/litellm/pull/18208) +### Logging -- **Pricing Updates** - - Correct pricing for `openrouter/openai/gpt-oss-20b` - [PR #18899](https://github.com/BerriAI/litellm/pull/18899) - - Add pricing for `azure_ai/claude-opus-4-5` - [PR #19003](https://github.com/BerriAI/litellm/pull/19003) - - Update Novita models prices - [PR #19005](https://github.com/BerriAI/litellm/pull/19005) - - Fix Azure Grok prices - [PR #19102](https://github.com/BerriAI/litellm/pull/19102) - - Fix GCP GLM-4.7 pricing - [PR #19172](https://github.com/BerriAI/litellm/pull/19172) - - Sync DeepSeek chat/reasoner to V3.2 pricing - [PR #18884](https://github.com/BerriAI/litellm/pull/18884) - - Correct cache_read pricing for gemini-2.5-pro models - [PR #18157](https://github.com/BerriAI/litellm/pull/18157) - -- **Budget & Rate Limiting** - - Correct budget limit validation operator (>=) for team members - [PR #19207](https://github.com/BerriAI/litellm/pull/19207) - - Fix TPM 25% limiting by ensuring priority queue logic - [PR #19092](https://github.com/BerriAI/litellm/pull/19092) - - Cleanup spend logs cron verification, fix, and docs - [PR #19085](https://github.com/BerriAI/litellm/pull/19085) +- **General Logging** + - prevent printing duplicate StandardLoggingPayload logs - [PR #19325](https://github.com/BerriAI/litellm/pull/19325) + - Fix: log duplication when json_logs is enabled - [PR #19705](https://github.com/BerriAI/litellm/pull/19705) +- **Langfuse OTEL** + - ignore service logs and fix callback shadowing - [PR #19298](https://github.com/BerriAI/litellm/pull/19298) +- **Langfuse** + - Send litellm_trace_id - [PR #19528](https://github.com/BerriAI/litellm/pull/19528) + - Add Langfuse mock mode for testing without API calls - [PR #19676](https://github.com/BerriAI/litellm/pull/19676) +- **GCS Bucket** + - prevent unbounded queue growth due to slow API calls - [PR #19297](https://github.com/BerriAI/litellm/pull/19297) + - Add GCS mock mode for testing without API calls - [PR #19683](https://github.com/BerriAI/litellm/pull/19683) +- **Responses API Logging** + - Fix pydantic serialization error - [PR #19486](https://github.com/BerriAI/litellm/pull/19486) +- **Arize Phoenix** + - add openinference span kinds to arize phoenix - [PR #19267](https://github.com/BerriAI/litellm/pull/19267) +- **Prometheus** + - Added new prometheus metrics for user count and team count - [PR #19520](https://github.com/BerriAI/litellm/pull/19520) ---- +### Guardrails -## MCP Gateway +- **Bedrock Guardrails** + - Ensure post_call guardrail checks input+output - [PR #19151](https://github.com/BerriAI/litellm/pull/19151) +- **Prompt Security** + - fixing prompt-security's guardrail implementation - [PR #19374](https://github.com/BerriAI/litellm/pull/19374) +- **Presidio** + - Fixes crash in Presidio Guardrail when running in background threads (logging_hook) - [PR #19714](https://github.com/BerriAI/litellm/pull/19714) +- **Pillar Security** + - Migrate Pillar Security to Generic Guardrail API - [PR #19364](https://github.com/BerriAI/litellm/pull/19364) +- **Policy Engine** + - New LiteLLM Policy engine - create policies to manage guardrails, conditions - permissions per Key, Team - [PR #19612](https://github.com/BerriAI/litellm/pull/19612) +- **General** + - add case-insensitive support for guardrail mode and actions - [PR #19480](https://github.com/BerriAI/litellm/pull/19480) -- Prevent duplicate MCP reload scheduler registration - [PR #18934](https://github.com/BerriAI/litellm/pull/18934) -- Forward MCP extra headers case-insensitively - [PR #18940](https://github.com/BerriAI/litellm/pull/18940) -- Fix MCP REST auth checks - [PR #19051](https://github.com/BerriAI/litellm/pull/19051) -- Fix generating two telemetry events in responses - [PR #18938](https://github.com/BerriAI/litellm/pull/18938) -- Fix MCP chat completions - [PR #19129](https://github.com/BerriAI/litellm/pull/19129) +### Prompt Management ---- +- **General** + - fix prompt info lookup and delete using correct IDs - [PR #19358](https://github.com/BerriAI/litellm/pull/19358) -## Performance / Loadbalancing / Reliability improvements +### Secret Manager -- **Performance Improvements** - - Remove bottleneck causing high CPU usage & overhead under heavy load - [PR #19049](https://github.com/BerriAI/litellm/pull/19049) - - Add CI enforcement for O(1) operations in `_get_model_cost_key` to prevent performance regressions - [PR #19052](https://github.com/BerriAI/litellm/pull/19052) - - Fix Azure embeddings JSON parsing to prevent connection leaks and ensure proper router cooldown - [PR #19167](https://github.com/BerriAI/litellm/pull/19167) - - Do not fallback to token counter if `disable_token_counter` is enabled - [PR #19041](https://github.com/BerriAI/litellm/pull/19041) +- **AWS Secret Manager** + - ensure auto-rotation updates existing AWS secret instead of creating new one - [PR #19455](https://github.com/BerriAI/litellm/pull/19455) +- **Hashicorp Vault** + - Ensure key rotations work with Vault - [PR #19634](https://github.com/BerriAI/litellm/pull/19634) -- **Reliability** - - Add fallback endpoints support - [PR #19185](https://github.com/BerriAI/litellm/pull/19185) - - Fix stream_timeout parameter functionality - [PR #19191](https://github.com/BerriAI/litellm/pull/19191) - - Fix model matching priority in configuration - [PR #19012](https://github.com/BerriAI/litellm/pull/19012) - - Fix num_retries in litellm_params as per config - [PR #18975](https://github.com/BerriAI/litellm/pull/18975) - - Handle exceptions without response parameter - [PR #18919](https://github.com/BerriAI/litellm/pull/18919) +--- -- **Infrastructure** - - Add Custom CA certificates to boto3 clients - [PR #18942](https://github.com/BerriAI/litellm/pull/18942) - - Update boto3 to 1.40.15 and aioboto3 to 15.5.0 - [PR #19090](https://github.com/BerriAI/litellm/pull/19090) - - Make keepalive_timeout parameter work for Gunicorn - [PR #19087](https://github.com/BerriAI/litellm/pull/19087) +## Spend Tracking, Budgets and Rate Limiting -- **Helm Chart** - - Fix mount config.yaml as single file in Helm chart - [PR #19146](https://github.com/BerriAI/litellm/pull/19146) - - Sync Helm chart versioning with production standards and Docker versions - [PR #18868](https://github.com/BerriAI/litellm/pull/18868) +- **Pricing Updates** + - Add openai/dall-e base pricing entries - [PR #19133](https://github.com/BerriAI/litellm/pull/19133) + - Add `input_cost_per_video_per_second` in ModelInfoBase - [PR #19398](https://github.com/BerriAI/litellm/pull/19398) --- -## Database Changes +## Performance / Loadbalancing / Reliability improvements -### Schema Updates -| Table | Change Type | Description | PR | -| ----- | ----------- | ----------- | -- | -| `LiteLLM_ProxyModelTable` | New Columns | Added `created_at` and `updated_at` timestamp fields | [PR #18937](https://github.com/BerriAI/litellm/pull/18937) | +- **General** + - Fix date overflow/division by zero in proxy utils - [PR #19527](https://github.com/BerriAI/litellm/pull/19527) + - Fix in-flight request termination on SIGTERM when health-check runs in a separate process - [PR #19427](https://github.com/BerriAI/litellm/pull/19427) + - Fix Pass through routes to work with server root path - [PR #19383](https://github.com/BerriAI/litellm/pull/19383) + - Fix logging error for stop iteration - [PR #19649](https://github.com/BerriAI/litellm/pull/19649) + - prevent retrying 4xx client errors - [PR #19275](https://github.com/BerriAI/litellm/pull/19275) + - add better error handling for misconfig on health check - [PR #19441](https://github.com/BerriAI/litellm/pull/19441) + +- **Router** + - Fix Azure RPM calculation formula - [PR #19513](https://github.com/BerriAI/litellm/pull/19513) + - Persist scheduler request queue to redis - [PR #19304](https://github.com/BerriAI/litellm/pull/19304) + - pass search_tools to Router during DB-triggered initialization - [PR #19388](https://github.com/BerriAI/litellm/pull/19388) + - Fixed PromptCachingCache to correctly handle messages where cache_control is a sibling key of string content - [PR #19266](https://github.com/BerriAI/litellm/pull/19266) + +- **Memory Leaks/OOM** + - prevent OOM with nested $defs in tool schemas - [PR #19112](https://github.com/BerriAI/litellm/pull/19112) + - fix: HTTP client memory leaks in Presidio, OpenAI, and Gemini - [PR #19190](https://github.com/BerriAI/litellm/pull/19190) + +- **Non root** + - fix logfile and pidfile of supervisor for non root environment - [PR #17267](https://github.com/BerriAI/litellm/pull/17267) + - resolve Read-only file system error in non-root images - [PR #19449](https://github.com/BerriAI/litellm/pull/19449) + +- **Dockerfile** + - Redis Semantic Caching - add missing redisvl dependency to requirements.txt - [PR #19417](https://github.com/BerriAI/litellm/pull/19417) + - Bump OTEL versions to support a2a dependency - resolves modulenotfounderror for Microsoft Agents by @Harshit28j in #18991 + +- **DB** + - Handle PostgreSQL cached plan errors during rolling deployments - [PR #19424](https://github.com/BerriAI/litellm/pull/19424) + +- **Timeouts** + - Fix: total timeout is not respected - [PR #19389](https://github.com/BerriAI/litellm/pull/19389) + +- **SDK** + - Field-Existence Checks to Type Classes to Prevent Attribute Errors - [PR #18321](https://github.com/BerriAI/litellm/pull/18321) + - add google-cloud-aiplatform as optional dependency with clear error message - [PR #19437](https://github.com/BerriAI/litellm/pull/19437) + - Make grpc dependency optional - [PR #19447](https://github.com/BerriAI/litellm/pull/19447) + - Add support for retry policies - [PR #19645](https://github.com/BerriAI/litellm/pull/19645) + +- **Performance** + - Cut chat_completion latency by ~21% by reducing pre-call processing time - [PR #19535](https://github.com/BerriAI/litellm/pull/19535) + - Optimize strip_trailing_slash with O(1) index check - [PR #19679](https://github.com/BerriAI/litellm/pull/19679) + - Optimize use_custom_pricing_for_model with set intersection - [PR #19677](https://github.com/BerriAI/litellm/pull/19677) + - perf: skip pattern_router.route() for non-wildcard models - [PR #19664](https://github.com/BerriAI/litellm/pull/19664) + - perf: Add LRU caching to get_model_info for faster cost lookups - [PR #19606](https://github.com/BerriAI/litellm/pull/19606) --- -## Documentation Updates - -- Add LiteLLM architecture md doc - [PR #19057](https://github.com/BerriAI/litellm/pull/19057), [PR #19252](https://github.com/BerriAI/litellm/pull/19252) -- Add troubleshooting guide - [PR #19096](https://github.com/BerriAI/litellm/pull/19096), [PR #19097](https://github.com/BerriAI/litellm/pull/19097), [PR #19099](https://github.com/BerriAI/litellm/pull/19099) -- Add structured issue reporting guides for CPU and memory issues - [PR #19117](https://github.com/BerriAI/litellm/pull/19117) -- Add Redis requirement warning for high-traffic deployments - [PR #18892](https://github.com/BerriAI/litellm/pull/18892) -- Update load balancing and routing with enable_pre_call_checks - [PR #18888](https://github.com/BerriAI/litellm/pull/18888) -- Updated pass_through with guided param - [PR #18886](https://github.com/BerriAI/litellm/pull/18886) -- Update message content types link and add content types table - [PR #18209](https://github.com/BerriAI/litellm/pull/18209) -- Add Redis initialization with kwargs - [PR #19183](https://github.com/BerriAI/litellm/pull/19183) -- Improve documentation for routing LLM calls via SAP Gen AI Hub - [PR #19166](https://github.com/BerriAI/litellm/pull/19166) -- Deleted Keys and Teams docs - [PR #19291](https://github.com/BerriAI/litellm/pull/19291) -- Claude Code end user tracking guide - [PR #19176](https://github.com/BerriAI/litellm/pull/19176) -- Add MCP troubleshooting guide - [PR #19122](https://github.com/BerriAI/litellm/pull/19122) -- Add auth message UI documentation - [PR #19063](https://github.com/BerriAI/litellm/pull/19063) -- Add guide for mounting custom callbacks in Helm/K8s - [PR #19136](https://github.com/BerriAI/litellm/pull/19136) +## General Proxy Improvements ---- +### Doc Improvements + - new tutorial for adding MCPs to Cursor via LiteLLM - [PR #19317](https://github.com/BerriAI/litellm/pull/19317) + - fix vertex_region to vertex_location in Vertex AI pass-through docs - [PR #19380](https://github.com/BerriAI/litellm/pull/19380) + - clarify Gemini and Vertex AI model prefix in json file - [PR #19443](https://github.com/BerriAI/litellm/pull/19443) + - update Claude Code integration guides - [PR #19415](https://github.com/BerriAI/litellm/pull/19415) + - adjust opencode tutorial - [PR #19605](https://github.com/BerriAI/litellm/pull/19605) + - add spend-queue-troubleshooting docs - [PR #19659](https://github.com/BerriAI/litellm/pull/19659) + - docs: add litellm-enterprise requirement for managed files - [PR #19689](https://github.com/BerriAI/litellm/pull/19689) -## Bug Fixes +### Helm + - Add support for keda in helm chart - [PR #19337](https://github.com/BerriAI/litellm/pull/19337) + - sync Helm chart version with LiteLLM release version - [PR #19438](https://github.com/BerriAI/litellm/pull/19438) + - Enable PreStop hook configuration in values.yaml - [PR #19613](https://github.com/BerriAI/litellm/pull/19613) + +### General + - Add health check scripts and parallel execution support - [PR #19295](https://github.com/BerriAI/litellm/pull/19295) -- Fix Swagger UI path execute error with server_root_path in OpenAPI schema - [PR #18947](https://github.com/BerriAI/litellm/pull/18947) -- Normalize OpenAI SDK BaseModel choices/messages to avoid Pydantic serializer warnings - [PR #18972](https://github.com/BerriAI/litellm/pull/18972) -- Add contextual gap checks and word-form digits - [PR #18301](https://github.com/BerriAI/litellm/pull/18301) -- Clean up orphaned files from repository root - [PR #19150](https://github.com/BerriAI/litellm/pull/19150) -- Include proxy/prisma_migration.py in non-root - [PR #18971](https://github.com/BerriAI/litellm/pull/18971) -- Update prisma_migration.py - [PR #19083](https://github.com/BerriAI/litellm/pull/19083) --- ## New Contributors -* @yogeshwaran10 made their first contribution in [PR #18898](https://github.com/BerriAI/litellm/pull/18898) -* @theonlypal made their first contribution in [PR #18937](https://github.com/BerriAI/litellm/pull/18937) -* @jonmagic made their first contribution in [PR #18935](https://github.com/BerriAI/litellm/pull/18935) -* @houdataali made their first contribution in [PR #19025](https://github.com/BerriAI/litellm/pull/19025) -* @hummat made their first contribution in [PR #18972](https://github.com/BerriAI/litellm/pull/18972) -* @berkeyalciin made their first contribution in [PR #18966](https://github.com/BerriAI/litellm/pull/18966) -* @MateuszOssGit made their first contribution in [PR #18959](https://github.com/BerriAI/litellm/pull/18959) -* @xfan001 made their first contribution in [PR #18947](https://github.com/BerriAI/litellm/pull/18947) -* @nulone made their first contribution in [PR #18884](https://github.com/BerriAI/litellm/pull/18884) -* @debnil-mercor made their first contribution in [PR #18919](https://github.com/BerriAI/litellm/pull/18919) -* @hakhundov made their first contribution in [PR #17420](https://github.com/BerriAI/litellm/pull/17420) -* @rohanwinsor made their first contribution in [PR #19078](https://github.com/BerriAI/litellm/pull/19078) -* @pgolm made their first contribution in [PR #19020](https://github.com/BerriAI/litellm/pull/19020) -* @vikigenius made their first contribution in [PR #19148](https://github.com/BerriAI/litellm/pull/19148) -* @burnerburnerburnerman made their first contribution in [PR #19090](https://github.com/BerriAI/litellm/pull/19090) -* @yfge made their first contribution in [PR #19076](https://github.com/BerriAI/litellm/pull/19076) -* @danielnyari-seon made their first contribution in [PR #19083](https://github.com/BerriAI/litellm/pull/19083) -* @guilherme-segantini made their first contribution in [PR #19166](https://github.com/BerriAI/litellm/pull/19166) -* @jgreek made their first contribution in [PR #19147](https://github.com/BerriAI/litellm/pull/19147) -* @anand-kamble made their first contribution in [PR #19193](https://github.com/BerriAI/litellm/pull/19193) -* @neubig made their first contribution in [PR #19162](https://github.com/BerriAI/litellm/pull/19162) + +* @dushyantzz made their first contribution in [PR #19158](https://github.com/BerriAI/litellm/pull/19158) +* @obod-mpw made their first contribution in [PR #19133](https://github.com/BerriAI/litellm/pull/19133) +* @msexxeta made their first contribution in [PR #19030](https://github.com/BerriAI/litellm/pull/19030) +* @rsicart made their first contribution in [PR #19337](https://github.com/BerriAI/litellm/pull/19337) +* @cluebbehusen made their first contribution in [PR #19311](https://github.com/BerriAI/litellm/pull/19311) +* @Lucky-Lodhi2004 made their first contribution in [PR #19315](https://github.com/BerriAI/litellm/pull/19315) +* @binbandit made their first contribution in [PR #19324](https://github.com/BerriAI/litellm/pull/19324) +* @flex-myeonghyeon made their first contribution in [PR #19381](https://github.com/BerriAI/litellm/pull/19381) +* @Lrakotoson made their first contribution in [PR #18321](https://github.com/BerriAI/litellm/pull/18321) +* @bensi94 made their first contribution in [PR #18787](https://github.com/BerriAI/litellm/pull/18787) +* @victorigualada made their first contribution in [PR #19368](https://github.com/BerriAI/litellm/pull/19368) +* @VedantMadane made their first contribution in #19266 +* @stiyyagura0901 made their first contribution in #19276 +* @kamilio made their first contribution in [PR #19447](https://github.com/BerriAI/litellm/pull/19447) +* @jonathansampson made their first contribution in [PR #19433](https://github.com/BerriAI/litellm/pull/19433) +* @rynecarbone made their first contribution in [PR #19416](https://github.com/BerriAI/litellm/pull/19416) +* @jayy-77 made their first contribution in #19366 +* @davida-ps made their first contribution in [PR #19374](https://github.com/BerriAI/litellm/pull/19374) +* @joaodinissf made their first contribution in [PR #19506](https://github.com/BerriAI/litellm/pull/19506) +* @ecao310 made their first contribution in [PR #19520](https://github.com/BerriAI/litellm/pull/19520) +* @mpcusack-altos made their first contribution in [PR #19577](https://github.com/BerriAI/litellm/pull/19577) +* @milan-berri made their first contribution in [PR #19602](https://github.com/BerriAI/litellm/pull/19602) +* @xqe2011 made their first contribution in #19621 --- ## Full Changelog -**[View complete changelog on GitHub](https://github.com/BerriAI/litellm/compare/v1.80.15.rc.1...v1.81.0.rc.1)** +**[View complete changelog on GitHub](https://github.com/BerriAI/litellm/releases/tag/v1.81.3.rc)** From 92607a89e571268d738ae3f6ad9d54eb57f9709f Mon Sep 17 00:00:00 2001 From: Krrish Dholakia Date: Mon, 26 Jan 2026 20:10:57 -0800 Subject: [PATCH 3/3] feat: update model table --- .../release_notes/v1.81.3-stable/index.md | 31 ++++++++++++------- 1 file changed, 19 insertions(+), 12 deletions(-) diff --git a/docs/my-website/release_notes/v1.81.3-stable/index.md b/docs/my-website/release_notes/v1.81.3-stable/index.md index 6c3f945b9e6..accacd0abdf 100644 --- a/docs/my-website/release_notes/v1.81.3-stable/index.md +++ b/docs/my-website/release_notes/v1.81.3-stable/index.md @@ -53,13 +53,18 @@ pip install litellm==1.81.3.rc.2 ### New Model Support -| Provider | Model | Features | -| -------- | ----- | -------- | -| OpenAI | `gpt-5.2-codex` | Code generation | -| Azure | `azure/gpt-5.2-codex` | Code generation | -| Cerebras | `cerebras/zai-glm-4.7` | Reasoning, function calling | -| Replicate | All chat models | Full support for all Replicate chat models | - +| Provider | Model | Context Window | Input ($/1M tokens) | Output ($/1M tokens) | Deprecation Date | +| -------- | ----- | -------------- | ------------------- | -------------------- | ---------------- | +| OpenAI | `gpt-audio`, `gpt-audio-2025-08-28` | 128K | $32/1M audio tokens, $2.5/1M text tokens | $64/1M audio tokens, $10/1M text tokens | - | +| OpenAI | `gpt-audio-mini`, `gpt-audio-mini-2025-08-28` | 128K | $10/1M audio tokens, $0.6/1M text tokens | $20/1M audio tokens, $2.4/1M text tokens | - | +| Deepinfra, Vertex AI, Google AI Studio, OpenRouter, Vercel AI Gateway | `gemini-2.0-flash-001`, `gemini-2.0-flash` | - | - | - | 2026-03-31 | +| Groq | `openai/gpt-oss-120b` | 131K | 0.075/1M cache read | 0.6/1M output tokens | - | +| Groq | `groq/openai/gpt-oss-20b` | 131K | 0.0375/1M cache read, $0.075/1M text tokens | 0.3/1M output tokens | - | +| Vertex AI | `gemini-2.5-computer-use-preview-10-2025` | 128K | $1.25 | $10 | - | +| Azure AI | `claude-haiku-4-5` | $1.25/1M cache read, $2/1M cache read above 1 hr, $0.1/1M text tokens | $5/1M output tokens | - | +| Azure AI | `claude-sonnet-4-5` | $3.75/1M cache read, $6/1M cache read above 1 hr, $3/1M text tokens | $15/1M output tokens | - | +| Azure AI | `claude-opus-4-5` | $6.25/1M cache read, $10/1M cache read above 1 hr, $0.5/1M text tokens | $25/1M output tokens | - | +| Azure AI | `claude-opus-4-1` | $18.75/1M cache read, $30/1M cache read above 1 hr, $1.5/1M text tokens | $75/1M output tokens | - | ### Features @@ -70,6 +75,7 @@ pip install litellm==1.81.3.rc.2 - **[VertexAI](../../docs/providers/vertex)** - Docs - Google Workload Identity Federation (WIF) support - [PR #19320](https://github.com/BerriAI/litellm/pull/19320) + - **[Agentcore](../../docs/providers/bedrock_agentcore)** - Fixes streaming issues with AWS Bedrock AgentCore where responses would stop after the first chunk, particularly affecting OAuth-enabled agents - [PR #17141](https://github.com/BerriAI/litellm/pull/17141) @@ -87,7 +93,6 @@ pip install litellm==1.81.3.rc.2 - **Gemini([Vertex AI](../../docs/providers/vertex), [Google AI Studio](../../docs/providers/gemini))** - use responseJsonSchema for Gemini 2.0+ models - [PR #19314](https://github.com/BerriAI/litellm/pull/19314) - - Update Gemini 2.0 Flash deprecation dates to March 31, 2026 - [PR #19592](https://github.com/BerriAI/litellm/pull/19592) - **[Volcengine](../../docs/providers/volcano)** - Support Volcengine responses api - [PR #18508](https://github.com/BerriAI/litellm/pull/18508) @@ -111,6 +116,7 @@ pip install litellm==1.81.3.rc.2 - **[Anthropic](../../docs/providers/anthropic)** - Fix anthropic-beta sent client side being overridden instead of appended to - [PR #19343](https://github.com/BerriAI/litellm/pull/19343) - Filter out unsupported fields from JSON schema for Anthropic's output_format API - [PR #19482](https://github.com/BerriAI/litellm/pull/19482) + - **[Bedrock](../../docs/providers/bedrock)** - Expose stability models via /image_edits endpoint and ensure proper request transformation - [PR #19323](https://github.com/BerriAI/litellm/pull/19323) - Claude Code x Bedrock Invoke fails with advanced-tool-use-2025-11-20 - [PR #19373](https://github.com/BerriAI/litellm/pull/19373) @@ -119,20 +125,21 @@ pip install litellm==1.81.3.rc.2 - Fix request validation errors when using Claude 4 via bedrock invoke - [PR #19381](https://github.com/BerriAI/litellm/pull/19381) - Handle thinking with tool calls for Claude 4 models - [PR #19506](https://github.com/BerriAI/litellm/pull/19506) - correct streaming choice index for tool calls - [PR #19506](https://github.com/BerriAI/litellm/pull/19506) -- **[Groq](../../docs/providers/groq)** - - correct Groq gpt-oss pricing and add cache pricing - [PR #19311](https://github.com/BerriAI/litellm/pull/19311) + - **[Ollama](../../docs/providers/ollama)** - Fix tool call errors due with improved message extraction - [PR #19369](https://github.com/BerriAI/litellm/pull/19369) + - **[VertexAI](../../docs/providers/vertex)** - Removed optional vertex_count_tokens_location param before request is sent to vertex - [PR #19359](https://github.com/BerriAI/litellm/pull/19359) + - **Gemini([Vertex AI](../../docs/providers/vertex), [Google AI Studio](../../docs/providers/gemini))** - Supports setting media_resolution and fps parameters on each video file, when using Gemini video understanding - [PR #19273](https://github.com/BerriAI/litellm/pull/19273) - - Add gemini-2.5-computer-use-preview-10-2025 model for vertex ai provider - [PR #19397](https://github.com/BerriAI/litellm/pull/19397) - handle reasoning_effort as dict from OpenAI Agents SDK - [PR #19419](https://github.com/BerriAI/litellm/pull/19419) - add file content support in tool results - [PR #19416](https://github.com/BerriAI/litellm/pull/19416) + - **[Azure](../../docs/providers/azure_ai)** - Fix Azure AI costs for Anthropic models - [PR #19530](https://github.com/BerriAI/litellm/pull/19530) - - Anthropic models on Azure AI cache pricing - [PR #19532](https://github.com/BerriAI/litellm/pull/19532) + - **[Giga Chat](../../docs/providers/gigachat)** - Add tool choice mapping - [PR #19645](https://github.com/BerriAI/litellm/pull/19645) ---