|
| 1 | +import type { ModelInfo } from "../model.js" |
| 2 | + |
| 3 | +export type FireworksModelId = |
| 4 | + | "accounts/fireworks/models/kimi-k2-instruct" |
| 5 | + | "accounts/fireworks/models/qwen3-235b-a22b-instruct-2507" |
| 6 | + | "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct" |
| 7 | + | "accounts/fireworks/models/deepseek-r1-0528" |
| 8 | + | "accounts/fireworks/models/deepseek-v3" |
| 9 | + |
| 10 | +export const fireworksDefaultModelId: FireworksModelId = "accounts/fireworks/models/kimi-k2-instruct" |
| 11 | + |
| 12 | +export const fireworksModels = { |
| 13 | + "accounts/fireworks/models/kimi-k2-instruct": { |
| 14 | + maxTokens: 16384, |
| 15 | + contextWindow: 128000, |
| 16 | + supportsImages: false, |
| 17 | + supportsPromptCache: false, |
| 18 | + inputPrice: 0.6, |
| 19 | + outputPrice: 2.5, |
| 20 | + description: |
| 21 | + "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities.", |
| 22 | + }, |
| 23 | + "accounts/fireworks/models/qwen3-235b-a22b-instruct-2507": { |
| 24 | + maxTokens: 32768, |
| 25 | + contextWindow: 256000, |
| 26 | + supportsImages: false, |
| 27 | + supportsPromptCache: false, |
| 28 | + inputPrice: 0.22, |
| 29 | + outputPrice: 0.88, |
| 30 | + description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025.", |
| 31 | + }, |
| 32 | + "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct": { |
| 33 | + maxTokens: 32768, |
| 34 | + contextWindow: 256000, |
| 35 | + supportsImages: false, |
| 36 | + supportsPromptCache: false, |
| 37 | + inputPrice: 0.45, |
| 38 | + outputPrice: 1.8, |
| 39 | + description: "Qwen3's most agentic code model to date.", |
| 40 | + }, |
| 41 | + "accounts/fireworks/models/deepseek-r1-0528": { |
| 42 | + maxTokens: 20480, |
| 43 | + contextWindow: 160000, |
| 44 | + supportsImages: false, |
| 45 | + supportsPromptCache: false, |
| 46 | + inputPrice: 3, |
| 47 | + outputPrice: 8, |
| 48 | + description: |
| 49 | + "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us.", |
| 50 | + }, |
| 51 | + "accounts/fireworks/models/deepseek-v3": { |
| 52 | + maxTokens: 16384, |
| 53 | + contextWindow: 128000, |
| 54 | + supportsImages: false, |
| 55 | + supportsPromptCache: false, |
| 56 | + inputPrice: 0.9, |
| 57 | + outputPrice: 0.9, |
| 58 | + description: |
| 59 | + "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us.", |
| 60 | + }, |
| 61 | +} as const satisfies Record<string, ModelInfo> |
0 commit comments