v1.55.8
github-actions
released this
21 Dec 03:47
·
1 commit
to 70a9ea99f25ed107839fc6e8103e1a87db78eb50
since this release
What's Changed
- fix(proxy_server.py): pass model access groups to get_key/get_team mo… by @krrishdholakia in #7281
- Litellm security fixes by @krrishdholakia in #7282
- Added sambanova cloud models by @rodrigo-92 in #7187
- Re-add prompt caching based model filtering (route to previous model) by @krrishdholakia in #7299
- (Fix) deprecated Pydantic Config class with model_config BerriAI/li… by @ishaan-jaff in #7300
- (feat - proxy) Add
status_code
tolitellm_proxy_total_requests_metric_total
by @ishaan-jaff in #7293 - fix(hosted_vllm/transformation.py): return fake api key, if none give… by @krrishdholakia in #7301
- LiteLLM Minor Fixes & Improvements (2024/12/18) p1 by @krrishdholakia in #7295
- (feat proxy) v2 - model max budgets by @ishaan-jaff in #7302
- (proxy admin ui) - show Teams sorted by
Team Alias
by @ishaan-jaff in #7296 - (Refactor) use separate file for track_cost_callback by @ishaan-jaff in #7304
- o1 - add image param handling by @krrishdholakia in #7312
- (code quality) run ruff rule to ban unused imports by @ishaan-jaff in #7313
- [Bug Fix]: ImportError: cannot import name 'T' from 're' by @ishaan-jaff in #7314
- (code refactor) - Add
BaseRerankConfig
. UseBaseRerankConfig
forcohere/rerank
andazure_ai/rerank
by @ishaan-jaff in #7319 - (feat) add infinity rerank models by @ishaan-jaff in #7321
- Litellm dev 12 19 2024 p2 by @krrishdholakia in #7315
- Langfuse Prompt Management Support by @krrishdholakia in #7322
- Fix LiteLLM Fireworks AI Documentation by @jravi-fireworks in #7333
New Contributors
- @rodrigo-92 made their first contribution in #7187
- @jravi-fireworks made their first contribution in #7333
Full Changelog: v1.55.4...v1.55.8
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.55.8
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 220.0 | 237.6551034099362 | 6.125601230624555 | 0.0 | 1832 | 0 | 193.92061900009594 | 1182.1513959999947 |
Aggregated | Passed ✅ | 220.0 | 237.6551034099362 | 6.125601230624555 | 0.0 | 1832 | 0 | 193.92061900009594 | 1182.1513959999947 |