Feature/upgrade to v1.81.0 stable#52
Conversation
…mpletions [fix] mcp chat completions
…e_gen Add openrouter support for image/generation endpoints
allows rendering on models.litellm.ai
… from docs maintain just 1 place for tutorials
…bedding_header_forwarding_with_model_group
…put_array_of_tokens
easier to surface
…ests Stabilise mock tests
…ime_fix Fix model name from query param in realtime request
…19148) * Added ability to customize logfire base url through env var * Added test to check if env var is used correctly for logfire * Document the env var * Documented env var in config_settings.md
[Docs] Deleted Keys and Teams Docs
…violation-msg feat(panw_prisma_airs): add custom violation message support
[Infra] Fix UI Build
… Code (BerriAI#19294) * fix websearch_interception_converted_stream * test_websearch_interception_no_tool_call_streaming * FakeAnthropicMessagesStreamIterator * LITELLM_WEB_SEARCH_TOOL_NAME * fixes tools def for litellm web search * fixes FakeAnthropicMessagesStreamIterator * test_litellm_standard_websearch_tool * use new hook for modfying before any transfroms from litellm * init WebSearchInterceptionLogger + ARCHITECTURE * fix config.yaml * init doc for claude code web search * docs fix * doc fix * fix mypy linting
…orwarded down to target (BerriAI#19524) * test_vertex_passthrough_forwards_anthropic_beta_header * add_incoming_headers
…ed down to LLM API (BerriAI#19542) * fix ALLOWED_VERTEX_AI_PASSTHROUGH_HEADERS * test_vertex_passthrough_forwards_anthropic_beta_header * fix test_vertex_passthrough_forwards_anthropic_beta_header * test_vertex_passthrough_does_not_forward_litellm_auth_token * fix utils * Using Anthropic Beta Features on Vertex AI * test_forward_headers_from_request_x_pass_prefix
…n a separate process (BerriAI#19427)
patches preserved: - FREE_MODELS budget bypass system (auth_checks.py, db_spend_update_writer.py, proxy_track_cost_callback.py) - Vision model routing with fallback (router.py, litellm_pre_call_utils.py, _types.py) - DAU/WAU/MAU analytics with team filtering (networking.tsx, new_usage.tsx) - GCS bucket logging integration (integrations/gcs_bucket/) - Budget duration support - daily/weekly/monthly (schema.prisma fields: budget_duration, budget_reset_at) Upstream improvements integrated: - Organization and team member budget checks (auth_checks.py) - Security fixes (CVE patches, expired key leak prevention) - PostgreSQL cached plan fixes - SIGTERM handling improvements - Budget validation enhancements - MCP server management mode controls - Extended EntityType support (organization, customer, agent) - Enhanced entity usage export (CSV/JSON utilities) Conflict resolution summary: - 104 total conflicts resolved (19 code files, 85 UI build artifacts) - Critical backend files auto-merged successfully (schema.prisma, auth_checks.py, router.py, litellm_pre_call_utils.py, proxy_track_cost_callback.py) - Manual resolutions: _types.py, db_spend_update_writer.py, GCS Readme, UI TypeScript files - UI build artifacts accepted from upstream (will be regenerated on next build) Testing required: - FREE_MODELS bypass functionality - Vision model fallback routing - Budget duration enforcement (daily/weekly/monthly) - DAU/WAU/MAU analytics endpoints - Team-based filtering - GCS logging to buckets - Organization budget checks - Team member budget checks
There was a problem hiding this comment.
Pull request overview
This PR upgrades LiteLLM to v1.81.0 stable, introducing significant new provider support, API endpoints, and infrastructure improvements. The update adds 4 new LLM providers, 2 new API endpoints, extensive documentation updates, and enhanced deployment configurations.
Changes:
- Added support for 4 new providers (Apertis AI, Abliteration, Amazon Nova, and enhanced Azure features)
- Introduced 2 new API endpoints (/interactions and /v1/messages/count_tokens)
- Enhanced documentation with 100+ new guides, provider docs, and integration examples
- Improved Docker and Helm deployment configurations with security hardening
- Updated CI/CD pipelines and testing infrastructure
Reviewed changes
Copilot reviewed 140 out of 2393 changed files in this pull request and generated 6 comments.
Show a summary per file
| File | Description |
|---|---|
| docs/my-website/docs/providers/* | Added documentation for new providers (Apertis, Abliteration, Amazon Nova, Azure enhancements) |
| docs/my-website/docs/observability/* | Added integration docs for Qualifire, Levo AI, Focus, Generic API, updated Phoenix and Datadog |
| docs/my-website/docs/projects/* | Added documentation for OpenAI Agents SDK, Harbor, GraphRAG, Google ADK, Agent Lightning |
| docs/my-website/docs/*.md | Added new endpoint documentation for /interactions, /anthropic_count_tokens, code interpreter, agent permissions |
| docker/* | Updated Dockerfiles with Windows line ending conversion and security improvements |
| deploy/charts/litellm-helm/* | Enhanced Helm chart with ServiceMonitor, topology spread, lifecycle hooks, and version updates |
| .github/workflows/* | Updated CI workflows with new labeling, duplicate checking, and improved testing |
| ci_cd/* | Enhanced security scanning with Grype config and secret detection preparation |
| cookbook/* | Added Braintrust prompt wrapper and AI coding tool guides |
💡 Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.
| aws_session_name: "my-test-session" | ||
| aws_role_name: "arn:aws:iam::335785316107:role/litellm-github-unit-tests-circleci" | ||
| aws_web_identity_token: "oidc/circleci_v2/" | ||
| aws_web_identity_token: "oidc/example-provider/" |
There was a problem hiding this comment.
The line changed from 'circleci_v2/' to 'example-provider/', but the section header still refers to CircleCI. Either update line 112's header to match the generic example, or revert this change to use a CircleCI-specific path.
| aws_web_identity_token: "oidc/example-provider/" | |
| aws_web_identity_token: "oidc/circleci_v2/" |
|
|
||
| :::info | ||
|
|
||
| This documentation is for **AgentCore Agents** (agent runtimes). If you want to use AgentCore MCP servers, add them as you would any other MCP server. See the [MCP documentation](https://docs.litellm.ai/docs/mcp) for details. |
There was a problem hiding this comment.
The MCP documentation link should be relative using '../mcp' instead of the full path to maintain consistency with other documentation links and avoid breakage if the domain changes.
| This documentation is for **AgentCore Agents** (agent runtimes). If you want to use AgentCore MCP servers, add them as you would any other MCP server. See the [MCP documentation](https://docs.litellm.ai/docs/mcp) for details. | |
| This documentation is for **AgentCore Agents** (agent runtimes). If you want to use AgentCore MCP servers, add them as you would any other MCP server. See the [MCP documentation](../mcp) for details. |
| @@ -14,6 +14,7 @@ priority=1 | |||
| exitcodes=0 | |||
| stopasgroup=true | |||
| killasgroup=true | |||
There was a problem hiding this comment.
The new SUPERVISORD_STOPWAITSECS environment variable is referenced but not documented. Add a comment explaining its purpose and default value, or document it in the deployment guide.
| killasgroup=true | |
| killasgroup=true | |
| ; SUPERVISORD_STOPWAITSECS controls how many seconds supervisord waits for a | |
| ; graceful shutdown before sending SIGKILL. If not set, supervisord's default | |
| ; stopwaitsecs (10 seconds) applies. This value is shared by both main and health. |
| exitcodes=0 | ||
| stopasgroup=true | ||
| killasgroup=true | ||
| stopwaitsecs=%(ENV_SUPERVISORD_STOPWAITSECS)s |
There was a problem hiding this comment.
The new SUPERVISORD_STOPWAITSECS environment variable is referenced but not documented. Add a comment explaining its purpose and default value, or document it in the deployment guide.
| @@ -2,6 +2,7 @@ | |||
|
|
|||
| if [ "$SEPARATE_HEALTH_APP" = "1" ]; then | |||
| export LITELLM_ARGS="$@" | |||
There was a problem hiding this comment.
The default value of 3600 seconds (1 hour) should be documented. Consider adding a comment explaining why this long timeout is necessary for graceful shutdown.
| export LITELLM_ARGS="$@" | |
| export LITELLM_ARGS="$@" | |
| # Use a long default (1 hour) so supervisord has enough time to gracefully stop | |
| # long-running worker processes and in-flight requests during shutdown. |
| @@ -3,6 +3,7 @@ | |||
| # Declare variables to be passed into your templates. | |||
|
|
|||
| replicaCount: 1 | |||
There was a problem hiding this comment.
The new numWorkers parameter is commented out but lacks documentation. Add a comment explaining what it controls and when users should configure it.
| replicaCount: 1 | |
| replicaCount: 1 | |
| # Number of worker processes per pod for handling LiteLLM requests; adjust this on larger nodes or under high concurrency, or leave unset to use the default. |
…/upgrade-to-v1.81.0-stable
Title
Relevant issues
Pre-Submission checklist
Please complete all items before asking a LiteLLM maintainer to review your PR
tests/litellm/directory, Adding at least 1 test is a hard requirement - see detailsmake test-unitType
🆕 New Feature
🐛 Bug Fix
🧹 Refactoring
📖 Documentation
🚄 Infrastructure
✅ Test
Changes