Skip to content

Releases: BerriAI/litellm

v1.30.6

09 Mar 22:16
e10991e
Compare
Choose a tag to compare

What's Changed

  • [Docs] Deploying litellm - litellm, litellm-database, litellm with redis by @ishaan-jaff in #2423
  • feat(helm-chart): redis as cache managed by chart by @debdutdeb in #2420

New Contributors

Full Changelog: v1.30.5...v1.30.6

v1.30.5

09 Mar 08:05
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.30.4...v1.30.5

v1.30.4

09 Mar 06:29
Compare
Choose a tag to compare

1.Incognito Requests - Don't log anything - docs: https://docs.litellm.ai/docs/proxy/enterprise#incognito-requests---dont-log-anything

When no-log=True, the request will not be logged on any callbacks and there will be no server logs on litellm

import openai
client = openai.OpenAI(
    api_key="anything",            # proxy api-key
    base_url="http://0.0.0.0:8000" # litellm proxy 
)

response = client.chat.completions.create(
    model="gpt-3.5-turbo",
    messages = [
        {
            "role": "user",
            "content": "this is a test request, write a short poem"
        }
    ],
    extra_body={
        "no-log": True
    }
)

print(response)

2. Allow user to pass messages.name for claude-3, perplexity

Note: Before this pr - the two providers would raise errors with the name param

LiteLLM SDK

import litellm
response = litellm.completion(
  model="claude-3-opus-20240229", 
  messages = [
    {"role": "user", "content": "Hi gm!", "name": "ishaan"},
   ]
)

LiteLLM Proxy Server

import openai
client = openai.OpenAI(
    api_key="anything",
    base_url="http://0.0.0.0:8000"
)

response = client.chat.completions.create(
model="claude-3-opus-20240229"", 
messages = [
    {"role": "user", "content": "Hi gm!", "name": "ishaan"},
])

print(response)

3. If user is using run_gunicorn use cpu_count to select optimal num_workers

4. AzureOpenAI - Pass api_version to litellm proxy per request

Usage - sending a request to litellm proxy

from openai import AzureOpenAI

client = AzureOpenAI(
    api_key="dummy",
    # I want to use a specific api_version, other than default 2023-07-01-preview
    api_version="2023-05-15",
    # OpenAI Proxy Endpoint
    azure_endpoint="https://openai-proxy.domain.com"
    )

response = client.chat.completions.create(
    model="gpt-35-turbo-16k-qt",
    messages=[
        {"role": "user", "content": "Some content"}
    ],
)

What's Changed

New Contributors

Full Changelog: v1.30.3...v1.30.4

v1.30.3

08 Mar 16:41
Compare
Choose a tag to compare

Full Changelog: v1.30.2...v1.30.3

v1.30.2

08 Mar 05:14
Compare
Choose a tag to compare

🚀 LiteLLM Proxy - Proxy 100+ LLMs, Set Budgets and Auto-Scale with the LiteLLM CloudFormation Stack 👉Start here: https://docs.litellm.ai/docs/proxy/deploy#aws-cloud-formation-stack

⚡️ Load Balancing - View Metrics about selected deployments in server logs

🔎 Proxy view better debug prisma logs / slack alerts

📖 Docs: setting load balancing config https://docs.litellm.ai/docs/proxy/configs

⭐️ PR for using cross account ARN with Bedrock, Sagemaker: #2179

https://github.com/BerriAI/litellm/releases/tag/v1.30.2

What's Changed

Group 5746

Full Changelog: v1.30.1...v1.30.2

v1.30.1

07 Mar 17:07
Compare
Choose a tag to compare

docs(team_based_routing.md): add docs on team based routing by @krrishdholakia
fix(proxy_server.py): fix model alias map + add back testing by @krrishdholakia

Full Changelog: v1.30.0...v1.30.1

v1.30.0

07 Mar 06:31
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.29.7...v1.30.0

v1.29.7

07 Mar 05:02
Compare
Choose a tag to compare

⚡️LiteLLM Proxy 100+ LLMs, Track Number of Requests, Avg Latency Per Model Deployment

model_latency

🛠️ High Traffic Fixes - Fix for DB connection limit hits when model fallbacks occur

🚀 High Traffic Fixes - /embedding - bug "Dictionary changed size during iteration"

⚡️ High Traffic Fixes - Switch off --detailed_debug in default Dockerfile. Users will need to opt in to viewing --detailed_debug logs. (This led to a 5% decrease in avg latency across 1K concurrent calls)

📖 Docs - Fixes for /user/new on LiteLLM Proxy Swagger (show how to set tpm/rpm limits per user) https://docs.litellm.ai/docs/proxy/virtual_keys#usernew

⭐️ Admin UI - separate latency, num requests graphs for model deployments https://docs.litellm.ai/docs/proxy/ui

What's Changed

Full Changelog: v1.29.4...v1.29.7

v1.29.5

06 Mar 05:23
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.29.3...v1.29.5

v1.29.4

06 Mar 04:56
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.29.3...v1.29.4