Skip to content

Releases: BerriAI/litellm

v1.63.0.dev5

06 Mar 16:49
Compare
Choose a tag to compare

What's Changed

  • Return signature on bedrock converse thinking + Fix {} empty dictionary on streaming + thinking by @krrishdholakia in #9023
  • (Refactor) /v1/messages to follow simpler logic for Anthropic API spec by @ishaan-jaff in #9013

Full Changelog: v1.63.0-nightly...v1.63.0.dev5

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.63.0.dev5

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 278.42101090109276 6.116149255066882 0.0 1830 0 214.94648899999902 4750.29671599998
Aggregated Passed ✅ 250.0 278.42101090109276 6.116149255066882 0.0 1830 0 214.94648899999902 4750.29671599998

v1.63.0.dev1

06 Mar 16:20
Compare
Choose a tag to compare

Full Changelog: v1.63.0-nightly...v1.63.0.dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.63.0.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 190.0 209.86284151312142 6.250523763835477 0.0 1867 0 163.62763399996538 3461.6653150000047
Aggregated Passed ✅ 190.0 209.86284151312142 6.250523763835477 0.0 1867 0 163.62763399996538 3461.6653150000047

v1.63.0-nightly

06 Mar 05:07
f6535ae
Compare
Choose a tag to compare

What's Changed

v1.63.0 fixes Anthropic 'thinking' response on streaming to return the signature block. Github Issue

It also moves the response structure from signature_delta to signature to be the same as Anthropic. Anthropic Docs

Diff

"message": {
    ...
    "reasoning_content": "The capital of France is Paris.",
    "thinking_blocks": [
        {
            "type": "thinking",
            "thinking": "The capital of France is Paris.",
-            "signature_delta": "EqoBCkgIARABGAIiQL2UoU0b1OHYi+..." # 👈 OLD FORMAT
+            "signature": "EqoBCkgIARABGAIiQL2UoU0b1OHYi+..." # 👈 KEY CHANGE
        }
    ]
}

Full Changelog: v1.62.4-nightly...v1.63.0-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.63.0-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.63.0-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 272.1226933173393 6.127690671911355 0.0 1834 0 217.38513100001455 3752.371346000018
Aggregated Passed ✅ 250.0 272.1226933173393 6.127690671911355 0.0 1834 0 217.38513100001455 3752.371346000018

v1.62.4-nightly

05 Mar 23:55
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.62.1-nightly...v1.62.4-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.62.4-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 230.0 255.24015655585677 6.161171624266898 0.0 1844 0 200.43409900000597 1911.432934000004
Aggregated Passed ✅ 230.0 255.24015655585677 6.161171624266898 0.0 1844 0 200.43409900000597 1911.432934000004

v1.62.1-nightly

03 Mar 23:18
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.61.20.rc...v1.62.1-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.62.1-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 264.65819933898456 6.116726934503129 0.0033443012217075608 1829 1 88.08753300002081 3128.6442510000256
Aggregated Passed ✅ 250.0 264.65819933898456 6.116726934503129 0.0033443012217075608 1829 1 88.08753300002081 3128.6442510000256

v1.61.20.rc.dev1

01 Mar 19:45
Compare
Choose a tag to compare

Full Changelog: v1.61.20.rc...v1.61.20.rc.dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.20.rc.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 180.0 191.79059422653788 6.202260528600653 0.0 1854 0 161.13682500002824 1614.0894129999879
Aggregated Passed ✅ 180.0 191.79059422653788 6.202260528600653 0.0 1854 0 161.13682500002824 1614.0894129999879

v1.61.20-stable

02 Mar 00:36
Compare
Choose a tag to compare

Full Changelog: v1.61.20.rc.dev1...v1.61.20-stable

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.61.20-stable

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 180.0 196.04185046530006 6.374061708350503 0.0 1902 0 162.48068100003366 3938.7112709999883
Aggregated Passed ✅ 180.0 196.04185046530006 6.374061708350503 0.0 1902 0 162.48068100003366 3938.7112709999883

v1.61.20.dev1

28 Feb 22:26
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.61.20.rc...v1.61.20.dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.20.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 240.0 254.1835915962261 6.2026122151904675 0.0 1855 0 211.9885020000538 1218.0083719999857
Aggregated Passed ✅ 240.0 254.1835915962261 6.2026122151904675 0.0 1855 0 211.9885020000538 1218.0083719999857

v1.61.20.rc

28 Feb 04:17
1e7b9cf
Compare
Choose a tag to compare
v1.61.20.rc Pre-release
Pre-release

What's Changed

  • Add new gpt-4.5-preview model + other updates by @krrishdholakia in #8879
  • (fix) Anthropic pass through cost tracking by @ishaan-jaff in #8874
  • (Bug Fix) - Accurate token counting for /anthropic/ API Routes on LiteLLM Proxy by @ishaan-jaff in #8880
  • (fix) Pass through spend tracking - ensure custom_llm_provider is tracked for Vertex, Google AI Studio, Anthropic by @ishaan-jaff in #8882

Full Changelog: v1.61.13-stable...v1.61.20.rc

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.20.rc

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 271.87043339059113 6.110384289826154 0.003342660990058071 1828 1 218.3162040000184 3401.295629999993
Aggregated Passed ✅ 250.0 271.87043339059113 6.110384289826154 0.003342660990058071 1828 1 218.3162040000184 3401.295629999993

v1.61.20-nightly

27 Feb 21:43
Compare
Choose a tag to compare

What's Changed

  • vertex ai anthropic thinking param support + cost calculation for vertex_ai/claude-3-7-sonnet by @krrishdholakia in #8853

Full Changelog: v1.61.19-nightly...v1.61.20-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.20-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 274.35423830943404 6.203695089639108 0.0033443100213687914 1855 1 99.50650100000757 3744.788088000007
Aggregated Passed ✅ 250.0 274.35423830943404 6.203695089639108 0.0033443100213687914 1855 1 99.50650100000757 3744.788088000007