|
27 | 27 | { |
28 | 28 | "model_name": "anthropic/claude-opus-4.5", |
29 | 29 | "aliases": [ |
30 | | - "opus", |
31 | | - "opus4.5", |
32 | | - "claude-opus" |
| 30 | + "opus4.5" |
33 | 31 | ], |
34 | 32 | "context_window": 200000, |
35 | 33 | "max_output_tokens": 64000, |
|
41 | 39 | "description": "Claude Opus 4.5 - Anthropic's frontier reasoning model for complex software engineering and agentic workflows", |
42 | 40 | "intelligence_score": 18 |
43 | 41 | }, |
| 42 | + { |
| 43 | + "model_name": "anthropic/claude-opus-4.6", |
| 44 | + "aliases": [ |
| 45 | + "opus", |
| 46 | + "opus4.6", |
| 47 | + "claude-opus" |
| 48 | + ], |
| 49 | + "context_window": 1000000, |
| 50 | + "max_output_tokens": 128000, |
| 51 | + "supports_extended_thinking": false, |
| 52 | + "supports_json_mode": false, |
| 53 | + "supports_function_calling": false, |
| 54 | + "supports_images": true, |
| 55 | + "max_image_size_mb": 5.0, |
| 56 | + "description": "Claude Opus 4.6 - Anthropic's strongest model for coding, long-running professional tasks, and agentic workflows", |
| 57 | + "intelligence_score": 18 |
| 58 | + }, |
44 | 59 | { |
45 | 60 | "model_name": "anthropic/claude-sonnet-4.5", |
46 | 61 | "aliases": [ |
47 | | - "sonnet", |
48 | 62 | "sonnet4.5" |
49 | 63 | ], |
50 | 64 | "context_window": 200000, |
|
57 | 71 | "description": "Claude Sonnet 4.5 - High-performance model with exceptional reasoning and efficiency", |
58 | 72 | "intelligence_score": 12 |
59 | 73 | }, |
| 74 | + { |
| 75 | + "model_name": "anthropic/claude-sonnet-4.6", |
| 76 | + "aliases": [ |
| 77 | + "sonnet", |
| 78 | + "sonnet4.6" |
| 79 | + ], |
| 80 | + "context_window": 1000000, |
| 81 | + "max_output_tokens": 128000, |
| 82 | + "supports_extended_thinking": false, |
| 83 | + "supports_json_mode": false, |
| 84 | + "supports_function_calling": false, |
| 85 | + "supports_images": true, |
| 86 | + "max_image_size_mb": 5.0, |
| 87 | + "description": "Claude Sonnet 4.6 - Frontier Sonnet with coding, agents, and professional task performance", |
| 88 | + "intelligence_score": 13 |
| 89 | + }, |
60 | 90 | { |
61 | 91 | "model_name": "anthropic/claude-opus-4.1", |
62 | 92 | "aliases": [ |
|
104 | 134 | }, |
105 | 135 | { |
106 | 136 | "model_name": "google/gemini-3-pro-preview", |
| 137 | + "aliases": [ |
| 138 | + "gemini3.0", |
| 139 | + "gemini-3.0", |
| 140 | + "pro-openrouter" |
| 141 | + ], |
| 142 | + "context_window": 1048576, |
| 143 | + "max_output_tokens": 65536, |
| 144 | + "supports_extended_thinking": true, |
| 145 | + "supports_json_mode": true, |
| 146 | + "supports_function_calling": true, |
| 147 | + "supports_images": true, |
| 148 | + "max_image_size_mb": 20.0, |
| 149 | + "allow_code_generation": true, |
| 150 | + "description": "Google's Gemini 3.0 Pro via OpenRouter with vision", |
| 151 | + "intelligence_score": 17 |
| 152 | + }, |
| 153 | + { |
| 154 | + "model_name": "google/gemini-3.1-pro-preview", |
107 | 155 | "aliases": [ |
108 | 156 | "pro", |
109 | 157 | "gemini-pro", |
110 | 158 | "gemini", |
111 | 159 | "gemini3", |
112 | | - "pro-openrouter" |
| 160 | + "gemini3.1" |
113 | 161 | ], |
114 | 162 | "context_window": 1048576, |
115 | 163 | "max_output_tokens": 65536, |
|
119 | 167 | "supports_images": true, |
120 | 168 | "max_image_size_mb": 20.0, |
121 | 169 | "allow_code_generation": true, |
122 | | - "description": "Google's Gemini 3.0 Pro via OpenRouter with vision", |
123 | | - "intelligence_score": 18 |
| 170 | + "description": "Google's Gemini 3.1 Pro - Frontier reasoning with enhanced software engineering and agentic capabilities", |
| 171 | + "intelligence_score": 19 |
124 | 172 | }, |
125 | 173 | { |
126 | 174 | "model_name": "google/gemini-2.5-pro", |
|
171 | 219 | "description": "Mistral's largest model (text-only)", |
172 | 220 | "intelligence_score": 11 |
173 | 221 | }, |
174 | | - { |
175 | | - "model_name": "meta-llama/llama-3-70b", |
176 | | - "aliases": [ |
177 | | - "llama", |
178 | | - "llama3", |
179 | | - "llama3-70b", |
180 | | - "llama-70b", |
181 | | - "llama3-openrouter" |
182 | | - ], |
183 | | - "context_window": 8192, |
184 | | - "max_output_tokens": 8192, |
185 | | - "supports_extended_thinking": false, |
186 | | - "supports_json_mode": false, |
187 | | - "supports_function_calling": false, |
188 | | - "supports_images": false, |
189 | | - "max_image_size_mb": 0.0, |
190 | | - "description": "Meta's Llama 3 70B model (text-only)", |
191 | | - "intelligence_score": 9 |
192 | | - }, |
193 | 222 | { |
194 | 223 | "model_name": "deepseek/deepseek-r1-0528", |
195 | 224 | "aliases": [ |
|
208 | 237 | "description": "DeepSeek R1 with thinking mode - advanced reasoning capabilities (text-only)", |
209 | 238 | "intelligence_score": 15 |
210 | 239 | }, |
211 | | - { |
212 | | - "model_name": "perplexity/llama-3-sonar-large-32k-online", |
213 | | - "aliases": [ |
214 | | - "perplexity", |
215 | | - "sonar", |
216 | | - "perplexity-online" |
217 | | - ], |
218 | | - "context_window": 32768, |
219 | | - "max_output_tokens": 32768, |
220 | | - "supports_extended_thinking": false, |
221 | | - "supports_json_mode": false, |
222 | | - "supports_function_calling": false, |
223 | | - "supports_images": false, |
224 | | - "max_image_size_mb": 0.0, |
225 | | - "description": "Perplexity's online model with web search (text-only)", |
226 | | - "intelligence_score": 9 |
227 | | - }, |
228 | 240 | { |
229 | 241 | "model_name": "openai/o3", |
230 | 242 | "aliases": [ |
|
316 | 328 | { |
317 | 329 | "model_name": "openai/gpt-5", |
318 | 330 | "aliases": [ |
319 | | - "gpt5" |
| 331 | + "gpt-5.0", |
| 332 | + "gpt5.0" |
320 | 333 | ], |
321 | 334 | "context_window": 400000, |
322 | 335 | "max_output_tokens": 128000, |
|
327 | 340 | "max_image_size_mb": 20.0, |
328 | 341 | "supports_temperature": true, |
329 | 342 | "temperature_constraint": "range", |
330 | | - "description": "GPT-5 (400K context, 128K output) - Advanced model with reasoning support", |
| 343 | + "description": "GPT-5.0 (400K context, 128K output) - Advanced model with reasoning support", |
331 | 344 | "intelligence_score": 16 |
332 | 345 | }, |
333 | 346 | { |
334 | 347 | "model_name": "openai/gpt-5.2-pro", |
335 | 348 | "aliases": [ |
336 | 349 | "gpt5.2-pro", |
337 | | - "gpt5.2pro", |
338 | | - "gpt5pro" |
| 350 | + "gpt5.2pro" |
339 | 351 | ], |
340 | 352 | "context_window": 400000, |
341 | 353 | "max_output_tokens": 272000, |
|
352 | 364 | "description": "GPT-5.2 Pro - Advanced reasoning model with highest quality responses (text+image input, text output only)", |
353 | 365 | "intelligence_score": 18 |
354 | 366 | }, |
| 367 | + { |
| 368 | + "model_name": "openai/gpt-5.4-pro", |
| 369 | + "aliases": [ |
| 370 | + "gpt5.4-pro", |
| 371 | + "gpt5.4pro", |
| 372 | + "gpt5pro" |
| 373 | + ], |
| 374 | + "context_window": 1050000, |
| 375 | + "max_output_tokens": 128000, |
| 376 | + "supports_extended_thinking": true, |
| 377 | + "supports_json_mode": true, |
| 378 | + "supports_function_calling": true, |
| 379 | + "supports_images": true, |
| 380 | + "max_image_size_mb": 20.0, |
| 381 | + "supports_temperature": false, |
| 382 | + "temperature_constraint": "fixed", |
| 383 | + "use_openai_response_api": true, |
| 384 | + "default_reasoning_effort": "high", |
| 385 | + "allow_code_generation": true, |
| 386 | + "description": "GPT-5.4 Pro - OpenAI's most advanced model with enhanced reasoning and 1M context window", |
| 387 | + "intelligence_score": 19 |
| 388 | + }, |
| 389 | + { |
| 390 | + "model_name": "openai/gpt-5.4", |
| 391 | + "aliases": [ |
| 392 | + "gpt5", |
| 393 | + "gpt5.4", |
| 394 | + "gpt-5.4" |
| 395 | + ], |
| 396 | + "context_window": 1050000, |
| 397 | + "max_output_tokens": 128000, |
| 398 | + "supports_extended_thinking": true, |
| 399 | + "supports_json_mode": true, |
| 400 | + "supports_function_calling": true, |
| 401 | + "supports_images": true, |
| 402 | + "max_image_size_mb": 20.0, |
| 403 | + "supports_temperature": false, |
| 404 | + "temperature_constraint": "fixed", |
| 405 | + "default_reasoning_effort": "medium", |
| 406 | + "allow_code_generation": true, |
| 407 | + "description": "GPT-5.4 - OpenAI's unified frontier model (1M context, 128K output) combining Codex and GPT capabilities", |
| 408 | + "intelligence_score": 19 |
| 409 | + }, |
355 | 410 | { |
356 | 411 | "model_name": "openai/gpt-5-codex", |
357 | 412 | "aliases": [ |
358 | | - "codex", |
| 413 | + "codex-5.0", |
359 | 414 | "gpt5codex" |
360 | 415 | ], |
361 | 416 | "context_window": 400000, |
|
450 | 505 | "description": "GPT-5.1 Codex (400K context, 128K output) - Agentic coding specialization available through the Responses API", |
451 | 506 | "intelligence_score": 19 |
452 | 507 | }, |
| 508 | + { |
| 509 | + "model_name": "openai/gpt-5.3-codex", |
| 510 | + "aliases": [ |
| 511 | + "codex", |
| 512 | + "codex-5.3", |
| 513 | + "gpt5.3-codex" |
| 514 | + ], |
| 515 | + "context_window": 400000, |
| 516 | + "max_output_tokens": 128000, |
| 517 | + "supports_extended_thinking": true, |
| 518 | + "supports_json_mode": true, |
| 519 | + "supports_function_calling": true, |
| 520 | + "supports_images": true, |
| 521 | + "max_image_size_mb": 20.0, |
| 522 | + "supports_temperature": false, |
| 523 | + "temperature_constraint": "fixed", |
| 524 | + "use_openai_response_api": true, |
| 525 | + "default_reasoning_effort": "high", |
| 526 | + "allow_code_generation": true, |
| 527 | + "description": "GPT-5.3 Codex - OpenAI's most advanced agentic coding model with frontier software engineering performance", |
| 528 | + "intelligence_score": 19 |
| 529 | + }, |
453 | 530 | { |
454 | 531 | "model_name": "openai/gpt-5.1-codex-mini", |
455 | 532 | "aliases": [ |
|
507 | 584 | "temperature_constraint": "range", |
508 | 585 | "description": "xAI's Grok 4.1 Fast Reasoning via OpenRouter (2M context) with vision and advanced reasoning", |
509 | 586 | "intelligence_score": 15 |
| 587 | + }, |
| 588 | + { |
| 589 | + "model_name": "deepseek/deepseek-v3.2-exp", |
| 590 | + "aliases": [ |
| 591 | + "deepseek-v3", |
| 592 | + "deepseek-v3.2", |
| 593 | + "dsv3" |
| 594 | + ], |
| 595 | + "context_window": 163840, |
| 596 | + "max_output_tokens": 65536, |
| 597 | + "supports_extended_thinking": true, |
| 598 | + "supports_json_mode": true, |
| 599 | + "supports_function_calling": false, |
| 600 | + "supports_images": false, |
| 601 | + "max_image_size_mb": 0.0, |
| 602 | + "description": "DeepSeek V3.2 Experimental - Strong reasoning capabilities (text-only)", |
| 603 | + "intelligence_score": 16 |
| 604 | + }, |
| 605 | + { |
| 606 | + "model_name": "mistralai/devstral-2512", |
| 607 | + "aliases": [ |
| 608 | + "devstral" |
| 609 | + ], |
| 610 | + "context_window": 262144, |
| 611 | + "max_output_tokens": 32768, |
| 612 | + "supports_extended_thinking": false, |
| 613 | + "supports_json_mode": true, |
| 614 | + "supports_function_calling": true, |
| 615 | + "supports_images": false, |
| 616 | + "max_image_size_mb": 0.0, |
| 617 | + "description": "Devstral 2 - Mistral's 123B parameter model specialized for agentic coding and codebase exploration", |
| 618 | + "intelligence_score": 15 |
| 619 | + }, |
| 620 | + { |
| 621 | + "model_name": "qwen/qwen3.5-397b-a17b", |
| 622 | + "aliases": [ |
| 623 | + "qwen", |
| 624 | + "qwen3.5" |
| 625 | + ], |
| 626 | + "context_window": 262144, |
| 627 | + "max_output_tokens": 65536, |
| 628 | + "supports_extended_thinking": true, |
| 629 | + "supports_json_mode": true, |
| 630 | + "supports_function_calling": true, |
| 631 | + "supports_images": true, |
| 632 | + "max_image_size_mb": 20.0, |
| 633 | + "description": "Qwen 3.5 397B - Frontier reasoning model with vision, hybrid architecture (text+image+video input)", |
| 634 | + "intelligence_score": 16 |
| 635 | + }, |
| 636 | + { |
| 637 | + "model_name": "minimax/minimax-m2.5", |
| 638 | + "aliases": [ |
| 639 | + "minimax", |
| 640 | + "m2.5" |
| 641 | + ], |
| 642 | + "context_window": 196608, |
| 643 | + "max_output_tokens": 32768, |
| 644 | + "supports_extended_thinking": false, |
| 645 | + "supports_json_mode": true, |
| 646 | + "supports_function_calling": true, |
| 647 | + "supports_images": false, |
| 648 | + "max_image_size_mb": 0.0, |
| 649 | + "description": "MiniMax M2.5 - SWE-Bench 80.2%, optimized for agent workflows and real-world productivity (API allows up to 196K output)", |
| 650 | + "intelligence_score": 16 |
510 | 651 | } |
511 | 652 | ] |
512 | 653 | } |
0 commit comments