From ea4860abb5683d63548dea36f893b1bce03523bc Mon Sep 17 00:00:00 2001 From: Roo Code Bot Date: Tue, 15 Jul 2025 12:52:46 +0000 Subject: [PATCH] fix: increase max_tokens for moonshotai/kimi-k2-instruct on Groq to 16384 - Updated maxTokens from 8192 to 16384 for moonshotai/kimi-k2-instruct model - Fixes Groq API error: "max_tokens must be less than or equal to 16384" - Aligns with Groq API limits for this specific model Fixes #5739 --- packages/types/src/providers/groq.ts | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/packages/types/src/providers/groq.ts b/packages/types/src/providers/groq.ts index 99bf4be3d01..2eac1f954a3 100644 --- a/packages/types/src/providers/groq.ts +++ b/packages/types/src/providers/groq.ts @@ -89,7 +89,7 @@ export const groqModels = { description: "DeepSeek R1 Distill Llama 70B model, 128K context.", }, "moonshotai/kimi-k2-instruct": { - maxTokens: 8192, + maxTokens: 16384, contextWindow: 131072, supportsImages: false, supportsPromptCache: false,