What's Changed
* [FEAT]- OTEL log litellm request / response by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4076
* [Feat] Enterprise - Attribute Management changes to Users in Audit Logs by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4083
* [FEAT]- OTEL Log raw LLM request/response on OTEL by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4078
* fix(cost_calculator.py): fixes tgai unmapped model pricing by krrishdholakia in https://github.com/BerriAI/litellm/pull/4085
* fix(utils.py): improved predibase exception mapping by krrishdholakia in https://github.com/BerriAI/litellm/pull/4080
* [Fix] Litellm sdk - allow ChatCompletionMessageToolCall, and Function to be used as dict by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4086
* Update together ai pricing by krrishdholakia in https://github.com/BerriAI/litellm/pull/4087
* [Feature]: Proxy: Support API-Key header in addition to Authorization header by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4088
* docs - cache controls on `litellm python SDK` by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4099
* docs: add llmcord.py to side bar nav by jakobdylanc in https://github.com/BerriAI/litellm/pull/4101
* docs: fix llmcord.py side bar link by jakobdylanc in https://github.com/BerriAI/litellm/pull/4104
* [FEAT] - viewing spend report per customer / team by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4105
* feat - log Proxy Server auth errors on OTEL by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4103
* [Feat] Client Side Fallbacks by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4107
* Fix typos: Enterpise -> Enterprise by msabramo in https://github.com/BerriAI/litellm/pull/4110
* `assistants.md`: Remove extra trailing backslash by msabramo in https://github.com/BerriAI/litellm/pull/4112
* `assistants.md`: Add "Get a Thread" example by msabramo in https://github.com/BerriAI/litellm/pull/4114
* ui - Fix Test Key dropdown by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4108
* fix(bedrock_httpx.py): fix tool calling for anthropic bedrock calls w/ streaming by krrishdholakia in https://github.com/BerriAI/litellm/pull/4106
* fix(proxy_server.py): allow passing in a list of team members by krrishdholakia in https://github.com/BerriAI/litellm/pull/4084
* fix - show `model group` in Azure ContentPolicy exceptions by ishaan-jaff in https://github.com/BerriAI/litellm/pull/4116
Client Side Fallbacks: https://docs.litellm.ai/docs/proxy/reliability#test---client-side-fallbacks
![fallbacks py](https://github.com/BerriAI/litellm/assets/29436595/14433d7d-1575-4886-bc44-61ede51806b0)
**Full Changelog**: https://github.com/BerriAI/litellm/compare/v1.40.7...v1.40.8
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.8
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.8
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
| Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
| /chat/completions | Passed ✅ | 140.0 | 169.11120714803027 | 6.281005310183787 | 0.0 | 1878 | 0 | 114.50119100004486 | 1457.4686270000257 |
| Aggregated | Passed ✅ | 140.0 | 169.11120714803027 | 6.281005310183787 | 0.0 | 1878 | 0 | 114.50119100004486 | 1457.4686270000257 |