Releases: BerriAI/litellm
v1.52.8
What's Changed
- (chore) ci/cd fix - use correct
test_key_generate_prisma.py
by @ishaan-jaff in #6718 - Litellm key update fix by @krrishdholakia in #6710
- Update code blocks in huggingface.md by @Aiden-Jeon in #6737
- Doc fix for prefix support by @CamdenClark in #6734
- (Feat) Add support for storing virtual keys in AWS SecretManager by @ishaan-jaff in #6728
- LiteLLM Minor Fixes & Improvement (11/14/2024) by @krrishdholakia in #6730
New Contributors
- @Aiden-Jeon made their first contribution in #6737
- @CamdenClark made their first contribution in #6734
Full Changelog: v1.52.6...v1.52.8
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.8
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 270.0 | 298.55231204572533 | 6.139888957283805 | 0.0 | 1837 | 0 | 232.112771000061 | 1744.873116000008 |
Aggregated | Passed ✅ | 270.0 | 298.55231204572533 | 6.139888957283805 | 0.0 | 1837 | 0 | 232.112771000061 | 1744.873116000008 |
v1.52.5-stable
Full Changelog: v1.52.5.staging1...v1.52.5-stable
Docker image ghcr.io/berriai/litellm:litellm_stable_nov12-stable
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_nov12-stable
What's Changed
- Litellm dev 11 11 2024 by @krrishdholakia in #6693
fix(init.py): add 'watsonx_text' as mapped llm api route
fix(opentelemetry.py): fix passing parallel tool calls to otel
fix(init.py): update provider-model mapping to include all known provider-model mappings
feat(anthropic): support passing document in llm api call
docs(anthropic.md): add pdf anthropic call to docs + expose new 'supports_pdf_input' function
- Add docs to export logs to Laminar by @dinmukhamedm in #6674
- (Feat) Add langsmith key based logging by @ishaan-jaff in #6682
- (fix) OpenAI's optional messages[].name does not work with Mistral API by @ishaan-jaff in #6701
- (feat) add xAI on Admin UI by @ishaan-jaff in #6680
- (docs) add benchmarks on 1K RPS by @ishaan-jaff in #6704
- (feat) add cost tracking stable diffusion 3 on Bedrock by @ishaan-jaff in #6676
- fix raise correct error 404 when /key/info is called on non-existent key by @ishaan-jaff in #6653
- (feat) Add support for logging to GCS Buckets with folder paths by @ishaan-jaff in #6675
- (feat) add bedrock image gen async support by @ishaan-jaff in #6672
- (feat) Add Bedrock Stability.ai Stable Diffusion 3 Image Generation models by @ishaan-jaff in #6673
- (Feat) 273% improvement GCS Bucket Logger - use Batched Logging by @ishaan-jaff in #6679
- Litellm Minor Fixes & Improvements (11/08/2024) by @krrishdholakia in #6658
fix(deepseek/chat): convert content list to str
test(test_deepseek_completion.py): implement base llm unit tests
fix(router.py): support content policy violation fallbacks with default fallbacks
fix(opentelemetry.py): refactor to move otel imports behing flag
fix(opentelemtry.py): close span on success completion
fix(user_api_key_auth.py): allow user_role to default to none
- (pricing): Fix multiple mistakes in Claude pricing by @Manouchehri in #6666
New Contributors
- @dinmukhamedm made their first contribution in #6674
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.5-stable
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 250.0 | 288.0333965427629 | 6.0955375578428805 | 0.0 | 1824 | 0 | 215.17615800001977 | 3641.4951400000177 |
Aggregated | Passed ✅ | 250.0 | 288.0333965427629 | 6.0955375578428805 | 0.0 | 1824 | 0 | 215.17615800001977 | 3641.4951400000177 |
v1.52.6
What's Changed
- LiteLLM Minor Fixes & Improvements (11/12/2024) by @krrishdholakia in #6705
- (feat) helm hook to sync db schema by @ishaan-jaff in #6715
- (fix proxy redis) Add redis sentinel support by @ishaan-jaff in #6154
- Fix: Update gpt-4o costs to those of gpt-4o-2024-08-06 by @klieret in #6714
- (fix) using Anthropic
response_format={"type": "json_object"}
by @ishaan-jaff in #6721 - (feat) Add cost tracking for Azure Dall-e-3 Image Generation + use base class to ensure basic image generation tests pass by @ishaan-jaff in #6716
New Contributors
Full Changelog: v1.52.5...v1.52.6
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.6
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 240.0 | 266.21521040425523 | 6.127671245386762 | 0.0 | 1833 | 0 | 215.80195500001764 | 2902.9665340000292 |
Aggregated | Passed ✅ | 240.0 | 266.21521040425523 | 6.127671245386762 | 0.0 | 1833 | 0 | 215.80195500001764 | 2902.9665340000292 |
v1.52.5.staging1
Full Changelog: v1.52.5...v1.52.5.staging1
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.5.staging1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 260.0 | 290.351197702432 | 6.041456642946781 | 0.0 | 1808 | 0 | 225.4500810000195 | 3132.288078999977 |
Aggregated | Passed ✅ | 260.0 | 290.351197702432 | 6.041456642946781 | 0.0 | 1808 | 0 | 225.4500810000195 | 3132.288078999977 |
v1.52.6.dev1
What's Changed
- chore: comment for maritalk by @nobu007 in #6607
- Update gpt-4o-2024-08-06, and o1-preview, o1-mini models in model cost map by @emerzon in #6654
- (QOL improvement) add unit testing for all static_methods in litellm_logging.py by @ishaan-jaff in #6640
- (feat) log error class, function_name on prometheus service failure hook + only log DB related failures on DB service hook by @ishaan-jaff in #6650
Full Changelog: v1.52.1...v1.52.6.dev1
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.6.dev1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.6.dev1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 250.0 | 284.54861130679984 | 6.047368285253406 | 0.0 | 1809 | 0 | 224.15313200008313 | 1935.1971319999848 |
Aggregated | Passed ✅ | 250.0 | 284.54861130679984 | 6.047368285253406 | 0.0 | 1809 | 0 | 224.15313200008313 | 1935.1971319999848 |
v1.52.5
What's Changed
- Litellm dev 11 11 2024 by @krrishdholakia in #6693
- Add docs to export logs to Laminar by @dinmukhamedm in #6674
- (Feat) Add langsmith key based logging by @ishaan-jaff in #6682
- (fix) OpenAI's optional messages[].name does not work with Mistral API by @ishaan-jaff in #6701
- (feat) add xAI on Admin UI by @ishaan-jaff in #6680
- (docs) add benchmarks on 1K RPS by @ishaan-jaff in #6704
- (feat) add cost tracking stable diffusion 3 on Bedrock by @ishaan-jaff in #6676
- fix raise correct error 404 when /key/info is called on non-existent key by @ishaan-jaff in #6653
New Contributors
- @dinmukhamedm made their first contribution in #6674
Full Changelog: v1.52.4...v1.52.5
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.5
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 200.0 | 216.13288200000045 | 6.215294300193555 | 0.0 | 1859 | 0 | 166.97629999998753 | 1726.1806539999611 |
Aggregated | Passed ✅ | 200.0 | 216.13288200000045 | 6.215294300193555 | 0.0 | 1859 | 0 | 166.97629999998753 | 1726.1806539999611 |
v1.52.4
What's Changed
- (feat) Add support for logging to GCS Buckets with folder paths by @ishaan-jaff in #6675
- (feat) add bedrock image gen async support by @ishaan-jaff in #6672
- (feat) Add Bedrock Stability.ai Stable Diffusion 3 Image Generation models by @ishaan-jaff in #6673
- (Feat) 273% improvement GCS Bucket Logger - use Batched Logging by @ishaan-jaff in #6679
Full Changelog: v1.52.3...v1.52.4
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.4
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 260.0 | 290.15274785816086 | 6.102299282865334 | 0.0 | 1826 | 0 | 221.48416699997142 | 3998.8694860000464 |
Aggregated | Passed ✅ | 260.0 | 290.15274785816086 | 6.102299282865334 | 0.0 | 1826 | 0 | 221.48416699997142 | 3998.8694860000464 |
v1.52.3
What's Changed
- Litellm Minor Fixes & Improvements (11/08/2024) by @krrishdholakia in #6658
- (pricing): Fix multiple mistakes in Claude pricing by @Manouchehri in #6666
Full Changelog: v1.52.2...v1.52.3
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.3
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 210.0 | 236.59706194640916 | 6.234242656262243 | 0.0 | 1866 | 0 | 180.61705699994945 | 3424.5764140000006 |
Aggregated | Passed ✅ | 210.0 | 236.59706194640916 | 6.234242656262243 | 0.0 | 1866 | 0 | 180.61705699994945 | 3424.5764140000006 |
v1.52.2-dev1
Full Changelog: v1.52.3...v1.52.2-dev1
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.2-dev1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.2-dev1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 200.0 | 220.40195196940357 | 6.227773108800657 | 0.0 | 1863 | 0 | 180.672578000042 | 2967.1460419999676 |
Aggregated | Passed ✅ | 200.0 | 220.40195196940357 | 6.227773108800657 | 0.0 | 1863 | 0 | 180.672578000042 | 2967.1460419999676 |
v1.52.2
What's Changed
- chore: comment for maritalk by @nobu007 in #6607
- Update gpt-4o-2024-08-06, and o1-preview, o1-mini models in model cost map by @emerzon in #6654
- (QOL improvement) add unit testing for all static_methods in litellm_logging.py by @ishaan-jaff in #6640
- (feat) log error class, function_name on prometheus service failure hook + only log DB related failures on DB service hook by @ishaan-jaff in #6650
- Update several Azure AI models in model cost map by @emerzon in #6655
- ci(conftest.py): reset conftest.py for local_testing/ by @krrishdholakia in #6657
- Litellm dev 11 07 2024 by @krrishdholakia in #6649
New Contributors
Full Changelog: v1.52.1...v1.52.2
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.2
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 230.0 | 251.09411961031876 | 6.087114215107422 | 0.0 | 1822 | 0 | 198.72582000004968 | 1667.4085729999888 |
Aggregated | Passed ✅ | 230.0 | 251.09411961031876 | 6.087114215107422 | 0.0 | 1822 | 0 | 198.72582000004968 | 1667.4085729999888 |