What's Changed
* [Feat] add VertexAI `vertex_ai/text-embedding-004` , `vertex_ai/text-multilingual-embedding-002 ` by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4145
* Add IAM cred caching for OIDC flow by Manouchehri in https://github.com/BerriAI/litellm/pull/3712
* feat(util.py/azure.py): Add OIDC support when running LiteLLM on Azure + Azure Upstream caching by Manouchehri in https://github.com/BerriAI/litellm/pull/3861
* [Feat] Support `task_type`, `auto_truncate` params by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4152
* [Feat] support `dimensions` for vertex embeddings by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4149
* docs - run proxy on custom root path by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4154
* [Fix] `user` was inserted in Proxy Server embedding requests + added param mapping for mistral by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4156
* [Fix] Add ClarifAI support for LiteLLM Proxy by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4158
* [Admin UI] Fix error Internal Users see when using SSO by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4164
* [Fix] - Error selecting model provider from UI by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4166
* [UI] add Azure AI studio models on UI by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4167
* feat(vertex_httpx.py): Support Vertex AI system messages, JSON Schema, etc. by krrishdholakia in https://github.com/BerriAI/litellm/pull/4160
* Fix errors in the Vertex AI documentation by yamitzky in https://github.com/BerriAI/litellm/pull/4171
* feat(prometheus): add api_team_alias to exported labels by bcvanmeurs in https://github.com/BerriAI/litellm/pull/4169
New Contributors
* yamitzky made their first contribution in https://github.com/BerriAI/litellm/pull/4171
**Full Changelog**: https://github.com/BerriAI/litellm/compare/v1.40.9...v1.40.10
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.10
Don't want to maintain your internal proxy? get in touch ๐
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
| Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
| /chat/completions | Passed โ
| 140.0 | 172.37660025809805 | 6.297822628765798 | 0.0 | 1883 | 0 | 114.60945100003528 | 3651.5153230000124 |
| Aggregated | Passed โ
| 140.0 | 172.37660025809805 | 6.297822628765798 | 0.0 | 1883 | 0 | 114.60945100003528 | 3651.5153230000124 |
v1.40.9-stable
**Full Changelog**: https://github.com/BerriAI/litellm/compare/v1.40.9...v1.40.9-stable
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.9-stable
Don't want to maintain your internal proxy? get in touch ๐
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
| Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
| /chat/completions | Passed โ
| 120.0 | 145.47575085580996 | 6.442239890874918 | 0.0 | 1928 | 0 | 104.64309999997568 | 1708.8100789999885 |
| Aggregated | Passed โ
| 120.0 | 145.47575085580996 | 6.442239890874918 | 0.0 | 1928 | 0 | 104.64309999997568 | 1708.8100789999885 |