diff --git a/aider/resources/model-metadata.json b/aider/resources/model-metadata.json index 64c480e55ed..c8be95697bb 100644 --- a/aider/resources/model-metadata.json +++ b/aider/resources/model-metadata.json @@ -711,5 +711,102 @@ "together_ai/Qwen/Qwen3-235B-A22B-fp8-tput": { "input_cost_per_token": 0.0000002, "output_cost_per_token": 0.0000006, - } + }, + "openai/qwen-3.6-plus": { + "max_tokens": 8192, + "max_input_tokens": 131072, + "max_output_tokens": 8192, + "input_cost_per_token": 0.000000439, + "output_cost_per_token": 0.000002633, + "litellm_provider": "openai", + "mode": "chat", + "supports_function_calling": true, + "supports_tool_choice": true, + "supports_system_messages": true + }, + "openai/deepseek-v3": { + "max_tokens": 8192, + "max_input_tokens": 160000, + "max_output_tokens": 8192, + "input_cost_per_token": 0.000000810, + "output_cost_per_token": 0.000002295, + "litellm_provider": "openai", + "mode": "chat", + "supports_function_calling": true, + "supports_tool_choice": true, + "supports_system_messages": true + }, + "openai/deepseek-r1": { + "max_tokens": 8192, + "max_input_tokens": 64000, + "max_output_tokens": 8192, + "input_cost_per_token": 0.000000675, + "output_cost_per_token": 0.000002903, + "litellm_provider": "openai", + "mode": "chat", + "supports_function_calling": false, + "supports_tool_choice": false, + "supports_system_messages": true + }, + "openai/kimi-k2.5": { + "max_tokens": 8192, + "max_input_tokens": 262144, + "max_output_tokens": 8192, + "input_cost_per_token": 0.000000675, + "output_cost_per_token": 0.000003780, + "litellm_provider": "openai", + "mode": "chat", + "supports_function_calling": true, + "supports_tool_choice": true, + "supports_system_messages": true + }, + "openai/gemma-4-31b": { + "max_tokens": 8192, + "max_input_tokens": 128000, + "max_output_tokens": 8192, + "input_cost_per_token": 0.000000189, + "output_cost_per_token": 0.000000540, + "litellm_provider": "openai", + "mode": "chat", + "supports_function_calling": true, + "supports_tool_choice": true, + "supports_vision": true, + "supports_system_messages": true + }, + "openai/qwen-3-32b": { + "max_tokens": 8192, + "max_input_tokens": 131072, + "max_output_tokens": 8192, + "input_cost_per_token": 0.000000392, + "output_cost_per_token": 0.000000810, + "litellm_provider": "openai", + "mode": "chat", + "supports_function_calling": true, + "supports_tool_choice": true, + "supports_system_messages": true + }, + "openai/llama-3.3-70b": { + "max_tokens": 8192, + "max_input_tokens": 131072, + "max_output_tokens": 8192, + "input_cost_per_token": 0.000001188, + "output_cost_per_token": 0.000001188, + "litellm_provider": "openai", + "mode": "chat", + "supports_function_calling": true, + "supports_tool_choice": true, + "supports_system_messages": true + }, + "openai/minimax-m2.5": { + "max_tokens": 8192, + "max_input_tokens": 196608, + "max_output_tokens": 8192, + "input_cost_per_token": 0.000000405, + "output_cost_per_token": 0.000001620, + "litellm_provider": "openai", + "mode": "chat", + "supports_function_calling": true, + "supports_tool_choice": true, + "supports_system_messages": true + } } diff --git a/aider/resources/model-settings.yml b/aider/resources/model-settings.yml index 7524c68fd43..58441476d14 100644 --- a/aider/resources/model-settings.yml +++ b/aider/resources/model-settings.yml @@ -2957,3 +2957,60 @@ use_repo_map: true use_temperature: false accepts_settings: ["reasoning_effort"] + +# Kyma API models (OpenAI-compatible gateway, set --openai-api-base https://kymaapi.com/v1) +- name: openai/qwen-3.6-plus + edit_format: diff + use_repo_map: true + examples_as_sys_msg: true + extra_params: + max_tokens: 8192 + +- name: openai/deepseek-v3 + edit_format: diff + use_repo_map: true + examples_as_sys_msg: true + extra_params: + max_tokens: 8192 + +- name: openai/deepseek-r1 + edit_format: diff + use_repo_map: true + examples_as_sys_msg: true + extra_params: + max_tokens: 8192 + +- name: openai/kimi-k2.5 + edit_format: diff + use_repo_map: true + examples_as_sys_msg: true + extra_params: + max_tokens: 8192 + +- name: openai/gemma-4-31b + edit_format: diff + use_repo_map: true + examples_as_sys_msg: true + extra_params: + max_tokens: 8192 + +- name: openai/qwen-3-32b + edit_format: diff + use_repo_map: true + examples_as_sys_msg: true + extra_params: + max_tokens: 8192 + +- name: openai/llama-3.3-70b + edit_format: diff + use_repo_map: true + examples_as_sys_msg: true + extra_params: + max_tokens: 8192 + +- name: openai/minimax-m2.5 + edit_format: diff + use_repo_map: true + examples_as_sys_msg: true + extra_params: + max_tokens: 8192