v1.52.9
github-actions
released this
16 Nov 03:25
·
1 commit
to 73ccbc0f148e97f036cd660cc2da9ba6c661eac0
since this release
What's Changed
- (feat) add bedrock/stability.stable-image-ultra-v1:0 by @ishaan-jaff in #6723
- [Feature]: Stop swallowing up AzureOpenAi exception responses in litellm's implementation for a BadRequestError by @ishaan-jaff in #6745
- [Feature]: json_schema in response support for Anthropic by @ishaan-jaff in #6748
- fix: import audio check by @IamRash-7 in #6740
- (fix) Cost tracking for
vertex_ai/imagen3
by @ishaan-jaff in #6752 - (feat) Vertex AI - add support for fine tuned embedding models by @ishaan-jaff in #6749
- LiteLLM Minor Fixes & Improvements (11/13/2024) by @krrishdholakia in #6729
- feat - add us.llama 3.1 models by @ishaan-jaff in #6760
- (Feat) Add Vertex Model Garden llama 3.1 models by @ishaan-jaff in #6763
- (fix) Fix - don't allow
viewer
roles to create virtual keys by @ishaan-jaff in #6764 - (feat) Use
litellm/
prefix when storing virtual keys in AWS secret manager by @ishaan-jaff in #6765
New Contributors
- @IamRash-7 made their first contribution in #6740
Full Changelog: v1.52.8...v1.52.9
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.9
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Failed ❌ | 280.0 | 314.28547197285235 | 6.039371468840217 | 0.0 | 1805 | 0 | 226.56484299994872 | 2776.9337409999935 |
Aggregated | Failed ❌ | 280.0 | 314.28547197285235 | 6.039371468840217 | 0.0 | 1805 | 0 | 226.56484299994872 | 2776.9337409999935 |