What's Changed
* Azure OpenAI improvements - o3 native streaming, improved tool call + response format handling by krrishdholakia in https://github.com/BerriAI/litellm/pull/8292
* Fix edit team on ui by krrishdholakia in https://github.com/BerriAI/litellm/pull/8295
* Improve rpm check on keys by krrishdholakia in https://github.com/BerriAI/litellm/pull/8301
* docs: fix enterprise links by wagnerjt in https://github.com/BerriAI/litellm/pull/8294
* Add gemini-2.0-flash pricing + model info by krrishdholakia in https://github.com/BerriAI/litellm/pull/8303
* Add Arize Cookbook for Turning on LiteLLM Proxy by exiao in https://github.com/BerriAI/litellm/pull/8336
* Add aistudio GEMINI 2.0 to model_prices_and_context_window.json by dceluis in https://github.com/BerriAI/litellm/pull/8335
* Fix pricing for Gemini 2.0 Flash 001 by elabbarw in https://github.com/BerriAI/litellm/pull/8320
* [DOCS] Update local_debugging.md by rokbenko in https://github.com/BerriAI/litellm/pull/8308
* (Bug Fix - Langfuse) - fix for when model response has `choices=[]` by ishaan-jaff in https://github.com/BerriAI/litellm/pull/8339
* Fixed meta llama 3.3 key for Databricks API by anton164 in https://github.com/BerriAI/litellm/pull/8093
* fix(utils.py): handle key error in msg validation by krrishdholakia in https://github.com/BerriAI/litellm/pull/8325
* (bug fix router.py) - safely handle `choices=[]` on llm responses by ishaan-jaff in https://github.com/BerriAI/litellm/pull/8342
* (QA+UI) - e2e flow for adding assembly ai passthrough endpoints by ishaan-jaff in https://github.com/BerriAI/litellm/pull/8337
New Contributors
* exiao made their first contribution in https://github.com/BerriAI/litellm/pull/8336
* dceluis made their first contribution in https://github.com/BerriAI/litellm/pull/8335
* rokbenko made their first contribution in https://github.com/BerriAI/litellm/pull/8308
* anton164 made their first contribution in https://github.com/BerriAI/litellm/pull/8093
**Full Changelog**: https://github.com/BerriAI/litellm/compare/v1.60.5...v1.60.6
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.60.6
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
| Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
| /chat/completions | Passed ✅ | 200.0 | 217.05167674521235 | 6.288425886864887 | 0.0 | 1880 | 0 | 164.17646499996863 | 2306.284880000021 |
| Aggregated | Passed ✅ | 200.0 | 217.05167674521235 | 6.288425886864887 | 0.0 | 1880 | 0 | 164.17646499996863 | 2306.284880000021 |