v1.52.16
github-actions
released this
26 Nov 10:22
·
1 commit
to 8673f2541e88331ff576df36b21cb55ceecd0330
since this release
What's Changed
- feat - allow sending
tags
on vertex pass through requests by @ishaan-jaff in #6876 - (feat) Add support for using @google/generative-ai JS with LiteLLM Proxy by @ishaan-jaff in #6899
- (UI fix) UI does not reload when you login / open a new tab by @ishaan-jaff in #6909
- (fix) pass through endpoints - run logging async + use thread pool executor for sync logging callbacks by @ishaan-jaff in #6907
- (redis fix) - fix
AbstractConnection.__init__() got an unexpected keyword argument 'ssl'
by @ishaan-jaff in #6908 - (docs) Simplify
/vertex_ai/
pass through docs by @ishaan-jaff in #6910
Full Changelog: v1.52.15...v1.52.16
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.16
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 230.0 | 261.12838541230485 | 6.194778256251999 | 0.0 | 1853 | 0 | 206.21302299997524 | 2167.8605710000056 |
Aggregated | Passed ✅ | 230.0 | 261.12838541230485 | 6.194778256251999 | 0.0 | 1853 | 0 | 206.21302299997524 | 2167.8605710000056 |