|
11 | 11 |
|
12 | 12 |
|
13 | 13 | class OpenAICompletionsSolver(QuestionSolver): |
14 | | - def __init__(self, config=None, enable_tx=False, enable_cache=False, priority=25): |
15 | | - super().__init__(config, enable_tx=enable_tx, enable_cache=enable_cache, priority=priority) |
| 14 | + def __init__(self, config=None, |
| 15 | + translator: Optional[LanguageTranslator] = None, |
| 16 | + detector: Optional[LanguageDetector] = None, |
| 17 | + priority: int = 50, |
| 18 | + enable_tx: bool = False, |
| 19 | + enable_cache: bool = False, |
| 20 | + internal_lang: Optional[str] = None): |
| 21 | + super().__init__(config=config, translator=translator, |
| 22 | + detector=detector, priority=priority, |
| 23 | + enable_tx=enable_tx, enable_cache=enable_cache, |
| 24 | + internal_lang=internal_lang) |
16 | 25 | self.api_url = f"{self.config.get('api_url', 'https://api.openai.com/v1')}/completions" |
17 | 26 | self.engine = self.config.get("model", "text-davinci-002") # "ada" cheaper and faster, "davinci" better |
18 | 27 | self.stop_token = "<|im_end|>" |
@@ -76,13 +85,10 @@ def post_process_sentence(text: str) -> str: |
76 | 85 |
|
77 | 86 |
|
78 | 87 | class OpenAIChatCompletionsSolver(ChatMessageSolver): |
79 | | - enable_tx = False |
80 | | - priority = 25 |
81 | | - |
82 | 88 | def __init__(self, config=None, |
83 | 89 | translator: Optional[LanguageTranslator] = None, |
84 | 90 | detector: Optional[LanguageDetector] = None, |
85 | | - priority: int = 50, |
| 91 | + priority: int = 25, |
86 | 92 | enable_tx: bool = False, |
87 | 93 | enable_cache: bool = False, |
88 | 94 | internal_lang: Optional[str] = None): |
|
0 commit comments