|
1 | | -from agentlab.llm.chat_api import ( |
2 | | - AzureModelArgs, |
3 | | - OpenAIModelArgs, |
4 | | - OpenRouterModelArgs, |
5 | | - SelfHostedModelArgs, |
6 | | -) |
| 1 | +from agentlab.llm.chat_api import (AzureModelArgs, OpenAIModelArgs, |
| 2 | + OpenRouterModelArgs, SelfHostedModelArgs) |
7 | 3 |
|
8 | 4 | default_oss_llms_args = { |
9 | 5 | "n_retry_server": 4, |
|
17 | 13 | ] |
18 | 14 |
|
19 | 15 | CHAT_MODEL_ARGS_DICT = { |
| 16 | + "openai/o3-mini-2025-01-31": OpenAIModelArgs( |
| 17 | + model_name="o3-mini-2025-01-31", |
| 18 | + max_total_tokens=200_000, |
| 19 | + max_input_tokens=200_000, |
| 20 | + max_new_tokens=100_000, |
| 21 | + vision_support=False, |
| 22 | + ), |
20 | 23 | "openai/gpt-4o-mini-2024-07-18": OpenAIModelArgs( |
21 | 24 | model_name="gpt-4o-mini-2024-07-18", |
22 | 25 | max_total_tokens=128_000, |
|
113 | 116 | **default_oss_llms_args, |
114 | 117 | ), |
115 | 118 | # ---------------- OPENROUTER ----------------# |
| 119 | + "openrouter/deepseek/deepseek-r1": OpenRouterModelArgs( |
| 120 | + model_name="deepseek/deepseek-r1", |
| 121 | + max_total_tokens=128_000, |
| 122 | + max_input_tokens=100_000, |
| 123 | + max_new_tokens=128_000, |
| 124 | + temperature=1e-1, |
| 125 | + ), |
116 | 126 | "openrouter/meta-llama/llama-3.1-405b-instruct": OpenRouterModelArgs( |
117 | 127 | model_name="meta-llama/llama-3.1-405b-instruct", |
118 | 128 | max_total_tokens=128_000, |
|
0 commit comments