Skip to content

Releases: BerriAI/litellm

v1.61.19.dev1

27 Feb 20:25
Compare
Choose a tag to compare

What's Changed

  • vertex ai anthropic thinking param support + cost calculation for vertex_ai/claude-3-7-sonnet by @krrishdholakia in #8853

Full Changelog: v1.61.19-nightly...v1.61.19.dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.19.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 279.81049794793904 6.161760602979549 0.0 1844 0 217.23908899997468 4265.23701100001
Aggregated Passed ✅ 250.0 279.81049794793904 6.161760602979549 0.0 1844 0 217.23908899997468 4265.23701100001

v1.61.19-nightly

27 Feb 06:27
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.61.17-nightly...v1.61.19-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.19-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 190.0 197.9691385267413 6.185555293319119 0.003341737057438746 1851 1 53.4934159999807 967.1408859999815
Aggregated Passed ✅ 190.0 197.9691385267413 6.185555293319119 0.003341737057438746 1851 1 53.4934159999807 967.1408859999815

v1.61.17.dev1

27 Feb 01:31
Compare
Choose a tag to compare

What's Changed

  • fix(get_litellm_params.py): handle no-log being passed in via kwargs by @krrishdholakia in #8830
  • fix(o_series_transformation.py): fix optional param check for o-serie… by @krrishdholakia in #8787
  • chore: set ttlSecondsAfterFinished on the migration job in the litellm-helm chart by @ashwin153 in #8593

New Contributors

Full Changelog: v1.61.17-nightly...v1.61.17.dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.17.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 170.0 197.91492646891462 6.342832402753434 6.342832402753434 1898 1898 138.25161099998695 4268.093897000028
Aggregated Failed ❌ 170.0 197.91492646891462 6.342832402753434 6.342832402753434 1898 1898 138.25161099998695 4268.093897000028

v1.61.17-nightly

26 Feb 19:14
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.61.16-nightly...v1.61.17-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.17-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.17-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 110.0 132.37492176765105 6.343856763236461 6.343856763236461 1898 1898 93.89094700003398 3315.9179240000185
Aggregated Failed ❌ 110.0 132.37492176765105 6.343856763236461 6.343856763236461 1898 1898 93.89094700003398 3315.9179240000185

v1.61.16-nightly

25 Feb 07:27
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.61.15-nightly...v1.61.16-nightly

## Docker Run LiteLLM Proxy

```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.16-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 150.0 173.09079325414655 6.248259928705226 6.248259928705226 1869 1869 131.67032300003711 1529.8640780000028
Aggregated Failed ❌ 150.0 173.09079325414655 6.248259928705226 6.248259928705226 1869 1869 131.67032300003711 1529.8640780000028

v1.61.15-nightly

23 Feb 08:11
566d935
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.61.13-nightly...v1.61.15-nightly

## Docker Run LiteLLM Proxy

```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.15-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 140.0 149.21232461729608 6.452882828983183 6.452882828983183 1931 1931 114.22628599996187 662.278525000005
Aggregated Failed ❌ 140.0 149.21232461729608 6.452882828983183 6.452882828983183 1931 1931 114.22628599996187 662.278525000005

v1.61.13-stable

23 Feb 04:16
Compare
Choose a tag to compare

Full Changelog: v1.61.13-nightly...v1.61.13-stable

🚨 Known Issue:

  • DD Trace was enabled by default on litellm docker: #8788
  • Expect a patched v1.61.13-stable with the fix

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.61.13-stable

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 180.0 204.35323738803527 6.313934995711798 6.313934995711798 1889 1889 146.12962300003574 2180.2391240000247
Aggregated Failed ❌ 180.0 204.35323738803527 6.313934995711798 6.313934995711798 1889 1889 146.12962300003574 2180.2391240000247

v1.55.8-stable-patched

22 Feb 16:45
Compare
Choose a tag to compare

Full Changelog: v1.55.8-stable...v1.55.8-stable-patched

## Docker Run LiteLLM Proxy

```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm_stable_release_branch-v1.55.8-stable-patched

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 150.0 172.39167931961987 6.334948735889217 6.334948735889217 1896 1896 131.01931900001773 2316.6445349999663
Aggregated Failed ❌ 150.0 172.39167931961987 6.334948735889217 6.334948735889217 1896 1896 131.01931900001773 2316.6445349999663

v1.61.13.rc

21 Feb 05:26
Compare
Choose a tag to compare

What's Changed

  • LiteLLM Contributor PRs (02/18/2025). by @krrishdholakia in #8643
  • fix(utils.py): handle token counter error when invalid message passed in by @krrishdholakia in #8670
  • (Bug fix) - Cache Health not working when configured with prometheus service logger by @ishaan-jaff in #8687
  • (Redis fix) - use mget_non_atomic by @ishaan-jaff in #8682
  • (Observability) - Add more detailed dd tracing on Proxy Auth, Bedrock Auth by @ishaan-jaff in #8693
  • (Infra/DB) - Allow running older litellm version when out of sync with current state of DB by @ishaan-jaff in #8695

Full Changelog: v1.61.11-nightly...v1.61.13.rc

## Docker Run LiteLLM Proxy

```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.13.rc

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 150.0 182.1689693668056 6.422359014184563 6.422359014184563 1922 1922 131.7662689999679 3100.3508269999998
Aggregated Failed ❌ 150.0 182.1689693668056 6.422359014184563 6.422359014184563 1922 1922 131.7662689999679 3100.3508269999998

v1.61.13-nightly

21 Feb 04:39
Compare
Choose a tag to compare

What's Changed

  • LiteLLM Contributor PRs (02/18/2025). by @krrishdholakia in #8643
  • fix(utils.py): handle token counter error when invalid message passed in by @krrishdholakia in #8670
  • (Bug fix) - Cache Health not working when configured with prometheus service logger by @ishaan-jaff in #8687
  • (Redis fix) - use mget_non_atomic by @ishaan-jaff in #8682
  • (Observability) - Add more detailed dd tracing on Proxy Auth, Bedrock Auth by @ishaan-jaff in #8693
  • (Infra/DB) - Allow running older litellm version when out of sync with current state of DB by @ishaan-jaff in #8695

Full Changelog: v1.61.11-nightly...v1.61.13-nightly

## Docker Run LiteLLM Proxy

```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.13-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 150.0 177.70403054473582 6.34954513279557 6.34954513279557 1900 1900 131.07585400001653 3605.6919650000054
Aggregated Failed ❌ 150.0 177.70403054473582 6.34954513279557 6.34954513279557 1900 1900 131.07585400001653 3605.6919650000054