22
33from __future__ import annotations
44
5+ import typing_extensions
56from typing import List , Union , Iterable
67from typing_extensions import Literal , overload
78
@@ -183,6 +184,7 @@ def batch_completion(
183184 cast_to = BatchCompletion ,
184185 )
185186
187+ @typing_extensions .deprecated ("chat_completion is deprecated. Please use /v1/openai/v1/chat/completions." )
186188 @overload
187189 def chat_completion (
188190 self ,
@@ -251,6 +253,7 @@ def chat_completion(
251253 """
252254 ...
253255
256+ @typing_extensions .deprecated ("chat_completion is deprecated. Please use /v1/openai/v1/chat/completions." )
254257 @overload
255258 def chat_completion (
256259 self ,
@@ -319,6 +322,7 @@ def chat_completion(
319322 """
320323 ...
321324
325+ @typing_extensions .deprecated ("chat_completion is deprecated. Please use /v1/openai/v1/chat/completions." )
322326 @overload
323327 def chat_completion (
324328 self ,
@@ -387,6 +391,7 @@ def chat_completion(
387391 """
388392 ...
389393
394+ @typing_extensions .deprecated ("chat_completion is deprecated. Please use /v1/openai/v1/chat/completions." )
390395 @required_args (["messages" , "model_id" ], ["messages" , "model_id" , "stream" ])
391396 def chat_completion (
392397 self ,
@@ -437,6 +442,7 @@ def chat_completion(
437442 stream_cls = Stream [ChatCompletionResponseStreamChunk ],
438443 )
439444
445+ @typing_extensions .deprecated ("/v1/inference/completion is deprecated. Please use /v1/openai/v1/completions." )
440446 @overload
441447 def completion (
442448 self ,
@@ -483,6 +489,7 @@ def completion(
483489 """
484490 ...
485491
492+ @typing_extensions .deprecated ("/v1/inference/completion is deprecated. Please use /v1/openai/v1/completions." )
486493 @overload
487494 def completion (
488495 self ,
@@ -529,6 +536,7 @@ def completion(
529536 """
530537 ...
531538
539+ @typing_extensions .deprecated ("/v1/inference/completion is deprecated. Please use /v1/openai/v1/completions." )
532540 @overload
533541 def completion (
534542 self ,
@@ -575,6 +583,7 @@ def completion(
575583 """
576584 ...
577585
586+ @typing_extensions .deprecated ("/v1/inference/completion is deprecated. Please use /v1/openai/v1/completions." )
578587 @required_args (["content" , "model_id" ], ["content" , "model_id" , "stream" ])
579588 def completion (
580589 self ,
@@ -617,6 +626,7 @@ def completion(
617626 stream_cls = Stream [CompletionResponse ],
618627 )
619628
629+ @typing_extensions .deprecated ("/v1/inference/embeddings is deprecated. Please use /v1/openai/v1/embeddings." )
620630 def embeddings (
621631 self ,
622632 * ,
@@ -821,6 +831,7 @@ async def batch_completion(
821831 cast_to = BatchCompletion ,
822832 )
823833
834+ @typing_extensions .deprecated ("chat_completion is deprecated. Please use /v1/openai/v1/chat/completions." )
824835 @overload
825836 async def chat_completion (
826837 self ,
@@ -889,6 +900,7 @@ async def chat_completion(
889900 """
890901 ...
891902
903+ @typing_extensions .deprecated ("chat_completion is deprecated. Please use /v1/openai/v1/chat/completions." )
892904 @overload
893905 async def chat_completion (
894906 self ,
@@ -957,6 +969,7 @@ async def chat_completion(
957969 """
958970 ...
959971
972+ @typing_extensions .deprecated ("chat_completion is deprecated. Please use /v1/openai/v1/chat/completions." )
960973 @overload
961974 async def chat_completion (
962975 self ,
@@ -1025,6 +1038,7 @@ async def chat_completion(
10251038 """
10261039 ...
10271040
1041+ @typing_extensions .deprecated ("chat_completion is deprecated. Please use /v1/openai/v1/chat/completions." )
10281042 @required_args (["messages" , "model_id" ], ["messages" , "model_id" , "stream" ])
10291043 async def chat_completion (
10301044 self ,
@@ -1075,6 +1089,7 @@ async def chat_completion(
10751089 stream_cls = AsyncStream [ChatCompletionResponseStreamChunk ],
10761090 )
10771091
1092+ @typing_extensions .deprecated ("/v1/inference/completion is deprecated. Please use /v1/openai/v1/completions." )
10781093 @overload
10791094 async def completion (
10801095 self ,
@@ -1121,6 +1136,7 @@ async def completion(
11211136 """
11221137 ...
11231138
1139+ @typing_extensions .deprecated ("/v1/inference/completion is deprecated. Please use /v1/openai/v1/completions." )
11241140 @overload
11251141 async def completion (
11261142 self ,
@@ -1167,6 +1183,7 @@ async def completion(
11671183 """
11681184 ...
11691185
1186+ @typing_extensions .deprecated ("/v1/inference/completion is deprecated. Please use /v1/openai/v1/completions." )
11701187 @overload
11711188 async def completion (
11721189 self ,
@@ -1213,6 +1230,7 @@ async def completion(
12131230 """
12141231 ...
12151232
1233+ @typing_extensions .deprecated ("/v1/inference/completion is deprecated. Please use /v1/openai/v1/completions." )
12161234 @required_args (["content" , "model_id" ], ["content" , "model_id" , "stream" ])
12171235 async def completion (
12181236 self ,
@@ -1255,6 +1273,7 @@ async def completion(
12551273 stream_cls = AsyncStream [CompletionResponse ],
12561274 )
12571275
1276+ @typing_extensions .deprecated ("/v1/inference/embeddings is deprecated. Please use /v1/openai/v1/embeddings." )
12581277 async def embeddings (
12591278 self ,
12601279 * ,
@@ -1327,14 +1346,20 @@ def __init__(self, inference: InferenceResource) -> None:
13271346 self .batch_completion = to_raw_response_wrapper (
13281347 inference .batch_completion ,
13291348 )
1330- self .chat_completion = to_raw_response_wrapper (
1331- inference .chat_completion ,
1349+ self .chat_completion = ( # pyright: ignore[reportDeprecated]
1350+ to_raw_response_wrapper (
1351+ inference .chat_completion # pyright: ignore[reportDeprecated],
1352+ )
13321353 )
1333- self .completion = to_raw_response_wrapper (
1334- inference .completion ,
1354+ self .completion = ( # pyright: ignore[reportDeprecated]
1355+ to_raw_response_wrapper (
1356+ inference .completion # pyright: ignore[reportDeprecated],
1357+ )
13351358 )
1336- self .embeddings = to_raw_response_wrapper (
1337- inference .embeddings ,
1359+ self .embeddings = ( # pyright: ignore[reportDeprecated]
1360+ to_raw_response_wrapper (
1361+ inference .embeddings # pyright: ignore[reportDeprecated],
1362+ )
13381363 )
13391364
13401365
@@ -1348,14 +1373,20 @@ def __init__(self, inference: AsyncInferenceResource) -> None:
13481373 self .batch_completion = async_to_raw_response_wrapper (
13491374 inference .batch_completion ,
13501375 )
1351- self .chat_completion = async_to_raw_response_wrapper (
1352- inference .chat_completion ,
1376+ self .chat_completion = ( # pyright: ignore[reportDeprecated]
1377+ async_to_raw_response_wrapper (
1378+ inference .chat_completion # pyright: ignore[reportDeprecated],
1379+ )
13531380 )
1354- self .completion = async_to_raw_response_wrapper (
1355- inference .completion ,
1381+ self .completion = ( # pyright: ignore[reportDeprecated]
1382+ async_to_raw_response_wrapper (
1383+ inference .completion # pyright: ignore[reportDeprecated],
1384+ )
13561385 )
1357- self .embeddings = async_to_raw_response_wrapper (
1358- inference .embeddings ,
1386+ self .embeddings = ( # pyright: ignore[reportDeprecated]
1387+ async_to_raw_response_wrapper (
1388+ inference .embeddings # pyright: ignore[reportDeprecated],
1389+ )
13591390 )
13601391
13611392
@@ -1369,14 +1400,20 @@ def __init__(self, inference: InferenceResource) -> None:
13691400 self .batch_completion = to_streamed_response_wrapper (
13701401 inference .batch_completion ,
13711402 )
1372- self .chat_completion = to_streamed_response_wrapper (
1373- inference .chat_completion ,
1403+ self .chat_completion = ( # pyright: ignore[reportDeprecated]
1404+ to_streamed_response_wrapper (
1405+ inference .chat_completion # pyright: ignore[reportDeprecated],
1406+ )
13741407 )
1375- self .completion = to_streamed_response_wrapper (
1376- inference .completion ,
1408+ self .completion = ( # pyright: ignore[reportDeprecated]
1409+ to_streamed_response_wrapper (
1410+ inference .completion # pyright: ignore[reportDeprecated],
1411+ )
13771412 )
1378- self .embeddings = to_streamed_response_wrapper (
1379- inference .embeddings ,
1413+ self .embeddings = ( # pyright: ignore[reportDeprecated]
1414+ to_streamed_response_wrapper (
1415+ inference .embeddings # pyright: ignore[reportDeprecated],
1416+ )
13801417 )
13811418
13821419
@@ -1390,12 +1427,18 @@ def __init__(self, inference: AsyncInferenceResource) -> None:
13901427 self .batch_completion = async_to_streamed_response_wrapper (
13911428 inference .batch_completion ,
13921429 )
1393- self .chat_completion = async_to_streamed_response_wrapper (
1394- inference .chat_completion ,
1430+ self .chat_completion = ( # pyright: ignore[reportDeprecated]
1431+ async_to_streamed_response_wrapper (
1432+ inference .chat_completion # pyright: ignore[reportDeprecated],
1433+ )
13951434 )
1396- self .completion = async_to_streamed_response_wrapper (
1397- inference .completion ,
1435+ self .completion = ( # pyright: ignore[reportDeprecated]
1436+ async_to_streamed_response_wrapper (
1437+ inference .completion # pyright: ignore[reportDeprecated],
1438+ )
13981439 )
1399- self .embeddings = async_to_streamed_response_wrapper (
1400- inference .embeddings ,
1440+ self .embeddings = ( # pyright: ignore[reportDeprecated]
1441+ async_to_streamed_response_wrapper (
1442+ inference .embeddings # pyright: ignore[reportDeprecated],
1443+ )
14011444 )
0 commit comments