v1.31.17
💵 Track LLM Spend per Team, Start here: https://docs.litellm.ai/docs/simple_proxy
🐳 New LiteLLM Helm Chart hosted on OCI GHCR: https://github.com/BerriAI/litellm/pkgs/container/litellm-helm
🚀 /health/readiness avg response time is now 93% faster
🛠️ Fix for /health/readiness when it returns large json as part of the success callback
📖 Docs on deploying LiteLLM with Helm Chart https://docs.litellm.ai/docs/proxy/deploy#quick-start
What's Changed
- Add function call result submission support for Claude 3 models by @lazyhope in #2527
- Update helm chart to accomodate recent project changes by @ShaunMaher in #2145
- [FEAT] LiteLLM Helm Chart hosted on ghcr + docs by @ishaan-jaff in #2551
- fix(proxy_server.py): write blocked user list to a db table by @krrishdholakia in #2552
- Litellm end user opt out v2 db by @krrishdholakia in #2554
- (feat) Proxy - improve health readiness perf (93% faster) by @ishaan-jaff in #2553
- (fix) /health/readiness return success callback names as (str) by @ishaan-jaff in #2557
- (fix) admin ui - order spend by date by @ishaan-jaff in #2559
New Contributors
Full Changelog: v1.31.16...v1.31.17
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 100.0 | 109.5266968426512 | 1.6136745626072975 | 0.0 | 483 | 0 | 94.7443700000008 | 693.4165350000399 |
/health/liveliness | Passed ✅ | 78 | 80.0886455294508 | 15.03089411422408 | 0.006681882246821108 | 4499 | 2 | 74.07101500001545 | 1331.0514590000082 |
/health/readiness | Passed ✅ | 78 | 79.42774233022519 | 15.408420461169474 | 0.0 | 4612 | 0 | 73.88309300000628 | 877.4356640000178 |
Aggregated | Passed ✅ | 78 | 81.2529662746507 | 32.052989138000854 | 0.006681882246821108 | 9594 | 2 | 73.88309300000628 | 1331.0514590000082 |