@@ -21,6 +21,7 @@ class TestCompletions:
2121 @parametrize
2222 def test_method_create_overload_1 (self , client : GradientAI ) -> None :
2323 completion = client .agents .chat .completions .create (
24+ agent_domain = "inference.do-ai.run" ,
2425 messages = [
2526 {
2627 "content" : "string" ,
@@ -35,6 +36,7 @@ def test_method_create_overload_1(self, client: GradientAI) -> None:
3536 @parametrize
3637 def test_method_create_with_all_params_overload_1 (self , client : GradientAI ) -> None :
3738 completion = client .agents .chat .completions .create (
39+ agent_domain = "inference.do-ai.run" ,
3840 messages = [
3941 {
4042 "content" : "string" ,
@@ -75,6 +77,7 @@ def test_method_create_with_all_params_overload_1(self, client: GradientAI) -> N
7577 @parametrize
7678 def test_raw_response_create_overload_1 (self , client : GradientAI ) -> None :
7779 response = client .agents .chat .completions .with_raw_response .create (
80+ agent_domain = "inference.do-ai.run" ,
7881 messages = [
7982 {
8083 "content" : "string" ,
@@ -93,6 +96,7 @@ def test_raw_response_create_overload_1(self, client: GradientAI) -> None:
9396 @parametrize
9497 def test_streaming_response_create_overload_1 (self , client : GradientAI ) -> None :
9598 with client .agents .chat .completions .with_streaming_response .create (
99+ agent_domain = "inference.do-ai.run" ,
96100 messages = [
97101 {
98102 "content" : "string" ,
@@ -113,6 +117,7 @@ def test_streaming_response_create_overload_1(self, client: GradientAI) -> None:
113117 @parametrize
114118 def test_method_create_overload_2 (self , client : GradientAI ) -> None :
115119 completion_stream = client .agents .chat .completions .create (
120+ agent_domain = "inference.do-ai.run" ,
116121 messages = [
117122 {
118123 "content" : "string" ,
@@ -128,6 +133,7 @@ def test_method_create_overload_2(self, client: GradientAI) -> None:
128133 @parametrize
129134 def test_method_create_with_all_params_overload_2 (self , client : GradientAI ) -> None :
130135 completion_stream = client .agents .chat .completions .create (
136+ agent_domain = "inference.do-ai.run" ,
131137 messages = [
132138 {
133139 "content" : "string" ,
@@ -168,6 +174,7 @@ def test_method_create_with_all_params_overload_2(self, client: GradientAI) -> N
168174 @parametrize
169175 def test_raw_response_create_overload_2 (self , client : GradientAI ) -> None :
170176 response = client .agents .chat .completions .with_raw_response .create (
177+ agent_domain = "inference.do-ai.run" ,
171178 messages = [
172179 {
173180 "content" : "string" ,
@@ -186,6 +193,7 @@ def test_raw_response_create_overload_2(self, client: GradientAI) -> None:
186193 @parametrize
187194 def test_streaming_response_create_overload_2 (self , client : GradientAI ) -> None :
188195 with client .agents .chat .completions .with_streaming_response .create (
196+ agent_domain = "inference.do-ai.run" ,
189197 messages = [
190198 {
191199 "content" : "string" ,
@@ -213,6 +221,7 @@ class TestAsyncCompletions:
213221 @parametrize
214222 async def test_method_create_overload_1 (self , async_client : AsyncGradientAI ) -> None :
215223 completion = await async_client .agents .chat .completions .create (
224+ agent_domain = "inference.do-ai.run" ,
216225 messages = [
217226 {
218227 "content" : "string" ,
@@ -227,6 +236,7 @@ async def test_method_create_overload_1(self, async_client: AsyncGradientAI) ->
227236 @parametrize
228237 async def test_method_create_with_all_params_overload_1 (self , async_client : AsyncGradientAI ) -> None :
229238 completion = await async_client .agents .chat .completions .create (
239+ agent_domain = "inference.do-ai.run" ,
230240 messages = [
231241 {
232242 "content" : "string" ,
@@ -267,6 +277,7 @@ async def test_method_create_with_all_params_overload_1(self, async_client: Asyn
267277 @parametrize
268278 async def test_raw_response_create_overload_1 (self , async_client : AsyncGradientAI ) -> None :
269279 response = await async_client .agents .chat .completions .with_raw_response .create (
280+ agent_domain = "inference.do-ai.run" ,
270281 messages = [
271282 {
272283 "content" : "string" ,
@@ -285,6 +296,7 @@ async def test_raw_response_create_overload_1(self, async_client: AsyncGradientA
285296 @parametrize
286297 async def test_streaming_response_create_overload_1 (self , async_client : AsyncGradientAI ) -> None :
287298 async with async_client .agents .chat .completions .with_streaming_response .create (
299+ agent_domain = "inference.do-ai.run" ,
288300 messages = [
289301 {
290302 "content" : "string" ,
@@ -305,6 +317,7 @@ async def test_streaming_response_create_overload_1(self, async_client: AsyncGra
305317 @parametrize
306318 async def test_method_create_overload_2 (self , async_client : AsyncGradientAI ) -> None :
307319 completion_stream = await async_client .agents .chat .completions .create (
320+ agent_domain = "inference.do-ai.run" ,
308321 messages = [
309322 {
310323 "content" : "string" ,
@@ -320,6 +333,7 @@ async def test_method_create_overload_2(self, async_client: AsyncGradientAI) ->
320333 @parametrize
321334 async def test_method_create_with_all_params_overload_2 (self , async_client : AsyncGradientAI ) -> None :
322335 completion_stream = await async_client .agents .chat .completions .create (
336+ agent_domain = "inference.do-ai.run" ,
323337 messages = [
324338 {
325339 "content" : "string" ,
@@ -360,6 +374,7 @@ async def test_method_create_with_all_params_overload_2(self, async_client: Asyn
360374 @parametrize
361375 async def test_raw_response_create_overload_2 (self , async_client : AsyncGradientAI ) -> None :
362376 response = await async_client .agents .chat .completions .with_raw_response .create (
377+ agent_domain = "inference.do-ai.run" ,
363378 messages = [
364379 {
365380 "content" : "string" ,
@@ -378,6 +393,7 @@ async def test_raw_response_create_overload_2(self, async_client: AsyncGradientA
378393 @parametrize
379394 async def test_streaming_response_create_overload_2 (self , async_client : AsyncGradientAI ) -> None :
380395 async with async_client .agents .chat .completions .with_streaming_response .create (
396+ agent_domain = "inference.do-ai.run" ,
381397 messages = [
382398 {
383399 "content" : "string" ,
0 commit comments