diff --git a/docs/my-website/docs/providers/azure.md b/docs/my-website/docs/providers/azure.md index 4895b2b1cfb1..bc141810cef7 100644 --- a/docs/my-website/docs/providers/azure.md +++ b/docs/my-website/docs/providers/azure.md @@ -74,6 +74,8 @@ response = litellm.completion( | gpt-4-32k | `completion('azure/', messages)` | | gpt-4-32k-0314 | `completion('azure/', messages)` | | gpt-4-32k-0613 | `completion('azure/', messages)` | +| gpt-4-1106-preview | `completion('azure/', messages)` | +| gpt-4-0125-preview | `completion('azure/', messages)` | | gpt-3.5-turbo | `completion('azure/', messages)` | | gpt-3.5-turbo-0301 | `completion('azure/', messages)` | | gpt-3.5-turbo-0613 | `completion('azure/', messages)` | diff --git a/litellm/model_prices_and_context_window_backup.json b/litellm/model_prices_and_context_window_backup.json index 67688f2dfbcd..e44f46ff84a0 100644 --- a/litellm/model_prices_and_context_window_backup.json +++ b/litellm/model_prices_and_context_window_backup.json @@ -252,6 +252,15 @@ "output_cost_per_pixel": 0.0, "litellm_provider": "openai" }, + "azure/gpt-4-0125-preview": { + "max_tokens": 128000, + "max_input_tokens": 128000, + "max_output_tokens": 4096, + "input_cost_per_token": 0.00001, + "output_cost_per_token": 0.00003, + "litellm_provider": "azure", + "mode": "chat" + }, "azure/gpt-4-1106-preview": { "max_tokens": 128000, "max_input_tokens": 128000, diff --git a/model_prices_and_context_window.json b/model_prices_and_context_window.json index 67688f2dfbcd..e44f46ff84a0 100644 --- a/model_prices_and_context_window.json +++ b/model_prices_and_context_window.json @@ -252,6 +252,15 @@ "output_cost_per_pixel": 0.0, "litellm_provider": "openai" }, + "azure/gpt-4-0125-preview": { + "max_tokens": 128000, + "max_input_tokens": 128000, + "max_output_tokens": 4096, + "input_cost_per_token": 0.00001, + "output_cost_per_token": 0.00003, + "litellm_provider": "azure", + "mode": "chat" + }, "azure/gpt-4-1106-preview": { "max_tokens": 128000, "max_input_tokens": 128000,