Skip to content

Releases: BerriAI/litellm

v1.18.4

19 Jan 05:04
79e261f
Compare
Choose a tag to compare

What's Changed

[Feat] Proxy - Add Spend tracking logs by @ishaan-jaff in #1498

New SpendTable when Using LiteLLM Virtual Keys - Logs API Key, CreatedAt Date + Time, Model, Spend, Messages, Response
Docs to get started: https://docs.litellm.ai/docs/proxy/virtual_keys

Group 197

[Feat] Proxy - Track Cost Per User (Using user passed to requests) by @ishaan-jaff in #1509

  • Proxy Server Track Cost Per User
    Request:
curl --location 'http://0.0.0.0:8000/chat/completions' \
        --header 'Content-Type: application/json' \
        --header 'Authorization: Bearer sk-RwPq' \
        --data ' {
        "model": "BEDROCK_GROUP",
        "user": "litellm-is-awesome-user",
        "messages": [
            {
            "role": "user",
            "content": "what llm are you-444"
            }
        ],
        }'

Cost Tracked in LiteLLM Spend Tracking DB

Screenshot 2024-01-18 at 5 56 17 PM

Notes:

  • If a user is passed to the request the proxy tracks cost for it
  • If the user does not exist in the User Table, we make a new user with the spend

feat(parallel_request_limiter.py): add support for tpm/rpm rate limits for keys by @krrishdholakia in #1501

Full Changelog: v1.18.3...v1.18.4

v1.18.3

19 Jan 01:34
Compare
Choose a tag to compare

What's Changed

curl 'http://0.0.0.0:8000/key/generate' \
--header 'Authorization: Bearer <your-master-key>' \
--header 'Content-Type: application/json' \
--data-raw '{
  "models": ["gpt-3.5-turbo", "gpt-4", "claude-2"],
  "team_id": "core-infra"
}'

Full Changelog: v1.18.2...v1.18.3

v1.18.2

18 Jan 21:03
Compare
Choose a tag to compare

What's Changed

  • [Test+Fix] /Key/Info, /Key/Update - Litellm unit test key endpoints by @ishaan-jaff in #1496
  • fix(ollama_chat.py): use tiktoken as backup for prompt token counting by @puffo in #1495
  • fix(parallel_request_limiter.py): decrement count for failed llm calls by @krrishdholakia in 1ea3833
  • fix(proxy_server.py): show all models user has access to in /models by @krrishdholakia in c8dd36d

New Contributors

Full Changelog: v1.18.1...v1.18.2

v1.18.1

18 Jan 17:54
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.18.0...v1.18.1

v1.18.0

18 Jan 02:55
Compare
Choose a tag to compare

What's Changed

https://docs.litellm.ai/docs/simple_proxy

  • [Feat] Proxy - Access Key metadata in callbacks by @ishaan-jaff in #1484
    • Access Proxy Key metadata in callbacks
    • Access Endpoint URL in calbacks - you can see if /chat/completions, /embeddings, /image/generation etc is called
    • Support for Langfuse Tags, We log request metadata as langfuse tags

PS. no keys leaked - these are keys to my local proxy
Screenshot 2024-01-17 at 6 10 10 PM

Support for model access groups

Use this if you have keys with access to specific models, and you want to give all them access to a new model.

You can now assign keys access to model groups, and add new models to that group via the config.yaml - https://docs.litellm.ai/docs/proxy/users#grant-access-to-new-model

curl --location 'http://localhost:8000/key/generate' \
-H 'Authorization: Bearer <your-master-key>' \
-H 'Content-Type: application/json' \
-d '{"models": ["beta-models"], # 👈 Model Access Group
            "max_budget": 0,}'

Langfuse Tags logged:

Screenshot 2024-01-17 at 6 11 36 PM * feat(proxy_server.py): support model access groups by @krrishdholakia in https://github.com//pull/1483

Full Changelog: v1.17.18...v1.18.0

What's Changed

Full Changelog: v1.17.18...v1.18.0

v1.17.18

18 Jan 01:39
Compare
Choose a tag to compare

What's Changed

  • [Fix+Test] /key/delete functions by @ishaan-jaff in #1482 Added extensive testing + improved swagger

Full Changelog: v1.17.17...v1.17.18

v1.17.17

17 Jan 22:03
Compare
Choose a tag to compare

What's Changed

Testing + fixes for: https://docs.litellm.ai/docs/proxy/virtual_keys

  1. Generate a Key, and use it to make a call
  2. Make a call with invalid key, expect it to fail
  3. Make a call to a key with invalid model - expect to fail
  4. Make a call to a key with valid model - expect to pass
  5. Make a call with key over budget, expect to fail
  6. Make a streaming chat/completions call with key over budget, expect to fail
  7. Make a call with an key that never expires, expect to pass
  8. Make a call with an expired key, expect to fail

Full Changelog: v1.17.16...v1.17.17

v1.17.16

17 Jan 20:39
Compare
Choose a tag to compare

Full Changelog: v1.17.15...v1.17.16

v1.17.15

17 Jan 19:50
Compare
Choose a tag to compare

What's Changed

Usage - with Azure Vision enhancements

Docs: https://docs.litellm.ai/docs/providers/azure#usage---with-azure-vision-enhancements

Note: Azure requires the base_url to be set with /extensions

Example

base_url=https://gpt-4-vision-resource.openai.azure.com/openai/deployments/gpt-4-vision/extensions
# base_url="{azure_endpoint}/openai/deployments/{azure_deployment}/extensions"

Usage

import os 
from litellm import completion

os.environ["AZURE_API_KEY"] = "your-api-key"

# azure call
response = completion(
            model="azure/gpt-4-vision",
            timeout=5,
            messages=[
                {
                    "role": "user",
                    "content": [
                        {"type": "text", "text": "Whats in this image?"},
                        {
                            "type": "image_url",
                            "image_url": {
                                "url": "https://avatars.githubusercontent.com/u/29436595?v=4"
                            },
                        },
                    ],
                }
            ],
            base_url="https://gpt-4-vision-resource.openai.azure.com/openai/deployments/gpt-4-vision/extensions",
            api_key=os.getenv("AZURE_VISION_API_KEY"),
            enhancements={"ocr": {"enabled": True}, "grounding": {"enabled": True}},
            dataSources=[
                {
                    "type": "AzureComputerVision",
                    "parameters": {
                        "endpoint": "https://gpt-4-vision-enhancement.cognitiveservices.azure.com/",
                        "key": os.environ["AZURE_VISION_ENHANCE_KEY"],
                    },
                }
            ],
)

Full Changelog: v1.17.14...v1.17.15

v1.17.14

17 Jan 18:06
Compare
Choose a tag to compare

Fixes bug for mistral ai api optional param mapping

Full Changelog: v1.17.13...v1.17.14