v1.56.3
版本发布时间: 2024-12-28 12:49:23
BerriAI/litellm最新发布版本:v1.56.9(2025-01-04 10:46:47)
What's Changed
- Update Documentation - Gemini Embedding by @igorlima in https://github.com/BerriAI/litellm/pull/7436
- (Bug fix) missing
model_group
field in logs for aspeech call types by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/7392 - (Feat) - new endpoint
GET /v1/fine_tuning/jobs/{fine_tuning_job_id:path}
by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/7427 - Update model_prices_and_context_window.json by @superpoussin22 in https://github.com/BerriAI/litellm/pull/7345
- LiteLLM Minor Fixes & Improvements (12/27/2024) - p1 by @krrishdholakia in https://github.com/BerriAI/litellm/pull/7448
- Litellm dev 12 27 2024 p2 1 by @krrishdholakia in https://github.com/BerriAI/litellm/pull/7449
New Contributors
- @igorlima made their first contribution in https://github.com/BerriAI/litellm/pull/7436
Full Changelog: https://github.com/BerriAI/litellm/compare/v1.56.2...v1.56.3
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.56.3
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 250.0 | 276.9724297749999 | 6.148940938190872 | 0.003341815727277648 | 1840 | 1 | 112.37049800001842 | 1700.1428350000083 |
Aggregated | Passed ✅ | 250.0 | 276.9724297749999 | 6.148940938190872 | 0.003341815727277648 | 1840 | 1 | 112.37049800001842 | 1700.1428350000083 |
1、 load_test.html 1.59MB
2、 load_test_stats.csv 596B