@@ -21,6 +21,7 @@ class TestCompletions:
2121 @parametrize
2222 def test_method_create_overload_1 (self , client : GradientAI ) -> None :
2323 completion = client .agents .chat .completions .create (
24+ agent_domain = "inference.do-ai.run" ,
2425 messages = [
2526 {
2627 "content" : "string" ,
@@ -35,6 +36,7 @@ def test_method_create_overload_1(self, client: GradientAI) -> None:
3536 @parametrize
3637 def test_method_create_with_all_params_overload_1 (self , client : GradientAI ) -> None :
3738 completion = client .agents .chat .completions .create (
39+ agent_domain = "inference.do-ai.run" ,
3840 messages = [
3941 {
4042 "content" : "string" ,
@@ -64,6 +66,7 @@ def test_method_create_with_all_params_overload_1(self, client: GradientAI) -> N
6466 @parametrize
6567 def test_raw_response_create_overload_1 (self , client : GradientAI ) -> None :
6668 response = client .agents .chat .completions .with_raw_response .create (
69+ agent_domain = "inference.do-ai.run" ,
6770 messages = [
6871 {
6972 "content" : "string" ,
@@ -82,6 +85,7 @@ def test_raw_response_create_overload_1(self, client: GradientAI) -> None:
8285 @parametrize
8386 def test_streaming_response_create_overload_1 (self , client : GradientAI ) -> None :
8487 with client .agents .chat .completions .with_streaming_response .create (
88+ agent_domain = "inference.do-ai.run" ,
8589 messages = [
8690 {
8791 "content" : "string" ,
@@ -102,6 +106,7 @@ def test_streaming_response_create_overload_1(self, client: GradientAI) -> None:
102106 @parametrize
103107 def test_method_create_overload_2 (self , client : GradientAI ) -> None :
104108 completion_stream = client .agents .chat .completions .create (
109+ agent_domain = "inference.do-ai.run" ,
105110 messages = [
106111 {
107112 "content" : "string" ,
@@ -117,6 +122,7 @@ def test_method_create_overload_2(self, client: GradientAI) -> None:
117122 @parametrize
118123 def test_method_create_with_all_params_overload_2 (self , client : GradientAI ) -> None :
119124 completion_stream = client .agents .chat .completions .create (
125+ agent_domain = "inference.do-ai.run" ,
120126 messages = [
121127 {
122128 "content" : "string" ,
@@ -146,6 +152,7 @@ def test_method_create_with_all_params_overload_2(self, client: GradientAI) -> N
146152 @parametrize
147153 def test_raw_response_create_overload_2 (self , client : GradientAI ) -> None :
148154 response = client .agents .chat .completions .with_raw_response .create (
155+ agent_domain = "inference.do-ai.run" ,
149156 messages = [
150157 {
151158 "content" : "string" ,
@@ -164,6 +171,7 @@ def test_raw_response_create_overload_2(self, client: GradientAI) -> None:
164171 @parametrize
165172 def test_streaming_response_create_overload_2 (self , client : GradientAI ) -> None :
166173 with client .agents .chat .completions .with_streaming_response .create (
174+ agent_domain = "inference.do-ai.run" ,
167175 messages = [
168176 {
169177 "content" : "string" ,
@@ -191,6 +199,7 @@ class TestAsyncCompletions:
191199 @parametrize
192200 async def test_method_create_overload_1 (self , async_client : AsyncGradientAI ) -> None :
193201 completion = await async_client .agents .chat .completions .create (
202+ agent_domain = "inference.do-ai.run" ,
194203 messages = [
195204 {
196205 "content" : "string" ,
@@ -205,6 +214,7 @@ async def test_method_create_overload_1(self, async_client: AsyncGradientAI) ->
205214 @parametrize
206215 async def test_method_create_with_all_params_overload_1 (self , async_client : AsyncGradientAI ) -> None :
207216 completion = await async_client .agents .chat .completions .create (
217+ agent_domain = "inference.do-ai.run" ,
208218 messages = [
209219 {
210220 "content" : "string" ,
@@ -234,6 +244,7 @@ async def test_method_create_with_all_params_overload_1(self, async_client: Asyn
234244 @parametrize
235245 async def test_raw_response_create_overload_1 (self , async_client : AsyncGradientAI ) -> None :
236246 response = await async_client .agents .chat .completions .with_raw_response .create (
247+ agent_domain = "inference.do-ai.run" ,
237248 messages = [
238249 {
239250 "content" : "string" ,
@@ -252,6 +263,7 @@ async def test_raw_response_create_overload_1(self, async_client: AsyncGradientA
252263 @parametrize
253264 async def test_streaming_response_create_overload_1 (self , async_client : AsyncGradientAI ) -> None :
254265 async with async_client .agents .chat .completions .with_streaming_response .create (
266+ agent_domain = "inference.do-ai.run" ,
255267 messages = [
256268 {
257269 "content" : "string" ,
@@ -272,6 +284,7 @@ async def test_streaming_response_create_overload_1(self, async_client: AsyncGra
272284 @parametrize
273285 async def test_method_create_overload_2 (self , async_client : AsyncGradientAI ) -> None :
274286 completion_stream = await async_client .agents .chat .completions .create (
287+ agent_domain = "inference.do-ai.run" ,
275288 messages = [
276289 {
277290 "content" : "string" ,
@@ -287,6 +300,7 @@ async def test_method_create_overload_2(self, async_client: AsyncGradientAI) ->
287300 @parametrize
288301 async def test_method_create_with_all_params_overload_2 (self , async_client : AsyncGradientAI ) -> None :
289302 completion_stream = await async_client .agents .chat .completions .create (
303+ agent_domain = "inference.do-ai.run" ,
290304 messages = [
291305 {
292306 "content" : "string" ,
@@ -316,6 +330,7 @@ async def test_method_create_with_all_params_overload_2(self, async_client: Asyn
316330 @parametrize
317331 async def test_raw_response_create_overload_2 (self , async_client : AsyncGradientAI ) -> None :
318332 response = await async_client .agents .chat .completions .with_raw_response .create (
333+ agent_domain = "inference.do-ai.run" ,
319334 messages = [
320335 {
321336 "content" : "string" ,
@@ -334,6 +349,7 @@ async def test_raw_response_create_overload_2(self, async_client: AsyncGradientA
334349 @parametrize
335350 async def test_streaming_response_create_overload_2 (self , async_client : AsyncGradientAI ) -> None :
336351 async with async_client .agents .chat .completions .with_streaming_response .create (
352+ agent_domain = "inference.do-ai.run" ,
337353 messages = [
338354 {
339355 "content" : "string" ,
0 commit comments