Skip to content

Releases: BerriAI/litellm

v1.52.2

08 Nov 15:12
1bef645
Compare
Choose a tag to compare

What's Changed

  • chore: comment for maritalk by @nobu007 in #6607
  • Update gpt-4o-2024-08-06, and o1-preview, o1-mini models in model cost map by @emerzon in #6654
  • (QOL improvement) add unit testing for all static_methods in litellm_logging.py by @ishaan-jaff in #6640
  • (feat) log error class, function_name on prometheus service failure hook + only log DB related failures on DB service hook by @ishaan-jaff in #6650
  • Update several Azure AI models in model cost map by @emerzon in #6655
  • ci(conftest.py): reset conftest.py for local_testing/ by @krrishdholakia in #6657
  • Litellm dev 11 07 2024 by @krrishdholakia in #6649

New Contributors

Full Changelog: v1.52.1...v1.52.2

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.2

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 230.0 251.09411961031876 6.087114215107422 0.0 1822 0 198.72582000004968 1667.4085729999888
Aggregated Passed ✅ 230.0 251.09411961031876 6.087114215107422 0.0 1822 0 198.72582000004968 1667.4085729999888

v1.52.1

07 Nov 20:53
27e1835
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.52.0...v1.52.1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 260.0 292.8286898309638 6.110969302244283 0.0 1828 0 230.12115400001676 2643.3588609999674
Aggregated Passed ✅ 260.0 292.8286898309638 6.110969302244283 0.0 1828 0 230.12115400001676 2643.3588609999674

v1.52.0-stable

09 Nov 02:05
695f48a
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.51.3...v1.52.0-stable

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.0-stable

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 240.0 270.29554346208295 6.124428891308869 0.0 1833 0 212.83106800001406 1622.2440090000418
Aggregated Passed ✅ 240.0 270.29554346208295 6.124428891308869 0.0 1833 0 212.83106800001406 1622.2440090000418

v1.52.0

05 Nov 18:16
695f48a
Compare
Choose a tag to compare

Group 6166

What's Changed

New Contributors

Full Changelog: v1.51.3...v1.52.0

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.0

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 210.0 231.0704959909717 6.291122309918775 0.0 1883 0 180.74613400000317 2618.3897060000163
Aggregated Passed ✅ 210.0 231.0704959909717 6.291122309918775 0.0 1883 0 180.74613400000317 2618.3897060000163

v1.51.3-dev1

04 Nov 11:17
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.51.3...v1.51.3-dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.51.3-dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 292.3714877928421 6.163980651581093 0.0 1844 0 226.11442700002726 2207.691740000001
Aggregated Passed ✅ 250.0 292.3714877928421 6.163980651581093 0.0 1844 0 226.11442700002726 2207.691740000001

v1.51.3

01 Nov 21:14
Compare
Choose a tag to compare

What's Changed

  • Support specifying exponential backoff retry strategy when calling completions() by @dbczumar in #6520
  • (fix) slack alerting - don't spam the failed cost tracking alert for the same model by @ishaan-jaff in #6543
  • (feat) add XAI ChatCompletion Support by @ishaan-jaff in #6373
  • LiteLLM Minor Fixes & Improvements (10/30/2024) by @krrishdholakia in #6519

Full Changelog: v1.51.2...v1.51.3

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.51.3

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 200.0 220.3819331893052 6.253936592654308 0.0 1870 0 179.7343989999831 3185.1700670000014
Aggregated Passed ✅ 200.0 220.3819331893052 6.253936592654308 0.0 1870 0 179.7343989999831 3185.1700670000014

v1.51.1-stable

01 Nov 17:43
5652c37
Compare
Choose a tag to compare

What's Changed

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_nov1-v1.51.1

Full Changelog: v1.51.2...v1.51.1-stable

custom-docker-image-test-dev

01 Nov 12:49
Compare
Choose a tag to compare

What's Changed

  • Support specifying exponential backoff retry strategy when calling completions() by @dbczumar in #6520

Full Changelog: v1.51.2...custom-docker-image-test-dev

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-custom-docker-image-test-dev

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-custom-docker-image-test-dev

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-custom-docker-image-test-dev

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-custom-docker-image-test-dev

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-custom-docker-image-test-dev

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 277.2980743547866 6.178135563258271 0.0 1849 0 222.02958399998352 3219.9342000000115
Aggregated Passed ✅ 250.0 277.2980743547866 6.178135563258271 0.0 1849 0 222.02958399998352 3219.9342000000115

v1.51.2

30 Oct 19:57
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.51.1...v1.51.2

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.51.2

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 287.73103211135674 6.039141910660648 0.0 1805 0 213.5724959999834 2930.6253560000073
Aggregated Passed ✅ 250.0 287.73103211135674 6.039141910660648 0.0 1805 0 213.5724959999834 2930.6253560000073

v1.51.1-staging

30 Oct 18:58
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.51.1...v1.51.1-staging

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.51.1-staging

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 270.0 311.93605914725106 6.080288332872121 0.0033408177653143525 1820 1 117.93499300000576 3293.080912999983
Aggregated Failed ❌ 270.0 311.93605914725106 6.080288332872121 0.0033408177653143525 1820 1 117.93499300000576 3293.080912999983