|
2 | 2 |
|
3 | 3 | from __future__ import annotations |
4 | 4 |
|
5 | | -from typing import Dict, Union, Iterable |
| 5 | +from typing import Dict, Iterable |
6 | 6 | from typing_extensions import Literal, overload |
7 | 7 |
|
8 | 8 | import httpx |
@@ -52,16 +52,7 @@ def create( |
52 | 52 | self, |
53 | 53 | *, |
54 | 54 | messages: Iterable[completion_create_params.Message], |
55 | | - model: Union[ |
56 | | - Literal[ |
57 | | - "Qwen/Qwen2.5-72B-Instruct-Turbo", |
58 | | - "Qwen/Qwen2.5-7B-Instruct-Turbo", |
59 | | - "Qwen/Qwen3.5-9B", |
60 | | - "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo", |
61 | | - "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo", |
62 | | - ], |
63 | | - str, |
64 | | - ], |
| 55 | + model: str, |
65 | 56 | chat_template_kwargs: object | Omit = omit, |
66 | 57 | compliance: Literal["hipaa"] | Omit = omit, |
67 | 58 | context_length_exceeded_behavior: Literal["truncate", "error"] | Omit = omit, |
@@ -207,16 +198,7 @@ def create( |
207 | 198 | self, |
208 | 199 | *, |
209 | 200 | messages: Iterable[completion_create_params.Message], |
210 | | - model: Union[ |
211 | | - Literal[ |
212 | | - "Qwen/Qwen2.5-72B-Instruct-Turbo", |
213 | | - "Qwen/Qwen2.5-7B-Instruct-Turbo", |
214 | | - "Qwen/Qwen3.5-9B", |
215 | | - "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo", |
216 | | - "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo", |
217 | | - ], |
218 | | - str, |
219 | | - ], |
| 201 | + model: str, |
220 | 202 | stream: Literal[True], |
221 | 203 | chat_template_kwargs: object | Omit = omit, |
222 | 204 | compliance: Literal["hipaa"] | Omit = omit, |
@@ -362,16 +344,7 @@ def create( |
362 | 344 | self, |
363 | 345 | *, |
364 | 346 | messages: Iterable[completion_create_params.Message], |
365 | | - model: Union[ |
366 | | - Literal[ |
367 | | - "Qwen/Qwen2.5-72B-Instruct-Turbo", |
368 | | - "Qwen/Qwen2.5-7B-Instruct-Turbo", |
369 | | - "Qwen/Qwen3.5-9B", |
370 | | - "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo", |
371 | | - "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo", |
372 | | - ], |
373 | | - str, |
374 | | - ], |
| 347 | + model: str, |
375 | 348 | stream: bool, |
376 | 349 | chat_template_kwargs: object | Omit = omit, |
377 | 350 | compliance: Literal["hipaa"] | Omit = omit, |
@@ -517,16 +490,7 @@ def create( |
517 | 490 | self, |
518 | 491 | *, |
519 | 492 | messages: Iterable[completion_create_params.Message], |
520 | | - model: Union[ |
521 | | - Literal[ |
522 | | - "Qwen/Qwen2.5-72B-Instruct-Turbo", |
523 | | - "Qwen/Qwen2.5-7B-Instruct-Turbo", |
524 | | - "Qwen/Qwen3.5-9B", |
525 | | - "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo", |
526 | | - "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo", |
527 | | - ], |
528 | | - str, |
529 | | - ], |
| 493 | + model: str, |
530 | 494 | chat_template_kwargs: object | Omit = omit, |
531 | 495 | compliance: Literal["hipaa"] | Omit = omit, |
532 | 496 | context_length_exceeded_behavior: Literal["truncate", "error"] | Omit = omit, |
@@ -629,16 +593,7 @@ async def create( |
629 | 593 | self, |
630 | 594 | *, |
631 | 595 | messages: Iterable[completion_create_params.Message], |
632 | | - model: Union[ |
633 | | - Literal[ |
634 | | - "Qwen/Qwen2.5-72B-Instruct-Turbo", |
635 | | - "Qwen/Qwen2.5-7B-Instruct-Turbo", |
636 | | - "Qwen/Qwen3.5-9B", |
637 | | - "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo", |
638 | | - "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo", |
639 | | - ], |
640 | | - str, |
641 | | - ], |
| 596 | + model: str, |
642 | 597 | chat_template_kwargs: object | Omit = omit, |
643 | 598 | compliance: Literal["hipaa"] | Omit = omit, |
644 | 599 | context_length_exceeded_behavior: Literal["truncate", "error"] | Omit = omit, |
@@ -784,16 +739,7 @@ async def create( |
784 | 739 | self, |
785 | 740 | *, |
786 | 741 | messages: Iterable[completion_create_params.Message], |
787 | | - model: Union[ |
788 | | - Literal[ |
789 | | - "Qwen/Qwen2.5-72B-Instruct-Turbo", |
790 | | - "Qwen/Qwen2.5-7B-Instruct-Turbo", |
791 | | - "Qwen/Qwen3.5-9B", |
792 | | - "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo", |
793 | | - "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo", |
794 | | - ], |
795 | | - str, |
796 | | - ], |
| 742 | + model: str, |
797 | 743 | stream: Literal[True], |
798 | 744 | chat_template_kwargs: object | Omit = omit, |
799 | 745 | compliance: Literal["hipaa"] | Omit = omit, |
@@ -939,16 +885,7 @@ async def create( |
939 | 885 | self, |
940 | 886 | *, |
941 | 887 | messages: Iterable[completion_create_params.Message], |
942 | | - model: Union[ |
943 | | - Literal[ |
944 | | - "Qwen/Qwen2.5-72B-Instruct-Turbo", |
945 | | - "Qwen/Qwen2.5-7B-Instruct-Turbo", |
946 | | - "Qwen/Qwen3.5-9B", |
947 | | - "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo", |
948 | | - "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo", |
949 | | - ], |
950 | | - str, |
951 | | - ], |
| 888 | + model: str, |
952 | 889 | stream: bool, |
953 | 890 | chat_template_kwargs: object | Omit = omit, |
954 | 891 | compliance: Literal["hipaa"] | Omit = omit, |
@@ -1094,16 +1031,7 @@ async def create( |
1094 | 1031 | self, |
1095 | 1032 | *, |
1096 | 1033 | messages: Iterable[completion_create_params.Message], |
1097 | | - model: Union[ |
1098 | | - Literal[ |
1099 | | - "Qwen/Qwen2.5-72B-Instruct-Turbo", |
1100 | | - "Qwen/Qwen2.5-7B-Instruct-Turbo", |
1101 | | - "Qwen/Qwen3.5-9B", |
1102 | | - "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo", |
1103 | | - "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo", |
1104 | | - ], |
1105 | | - str, |
1106 | | - ], |
| 1034 | + model: str, |
1107 | 1035 | chat_template_kwargs: object | Omit = omit, |
1108 | 1036 | compliance: Literal["hipaa"] | Omit = omit, |
1109 | 1037 | context_length_exceeded_behavior: Literal["truncate", "error"] | Omit = omit, |
|
0 commit comments