Skip to content

Conversation

@mrubens
Copy link
Collaborator

@mrubens mrubens commented Sep 5, 2025

Important

Add moonshotai/kimi-k2-instruct-0905 model to groq.ts and set it as the default model with specific configurations.

  • Models:
    • Add moonshotai/kimi-k2-instruct-0905 to GroqModelId and groqModels in groq.ts.
    • Set groqDefaultModelId to moonshotai/kimi-k2-instruct-0905 in groq.ts.
  • Configurations:
    • moonshotai/kimi-k2-instruct-0905 has maxTokens: 16384, contextWindow: 262144, supportsImages: false, supportsPromptCache: true, inputPrice: 0.6, outputPrice: 2.5, cacheReadsPrice: 0.15.

This description was created by Ellipsis for c43f39e. You can customize this summary. It will automatically update as commits are pushed.

@mrubens mrubens requested review from cte and jr as code owners September 5, 2025 03:43
@dosubot dosubot bot added size:S This PR changes 10-29 lines, ignoring generated files. enhancement New feature or request labels Sep 5, 2025
Copy link
Contributor

@roomote roomote bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thank you for your contribution! I've reviewed the changes and found several issues that need attention before this can be merged.

| "openai/gpt-oss-20b"

export const groqDefaultModelId: GroqModelId = "llama-3.3-70b-versatile" // Defaulting to Llama3 70B Versatile
export const groqDefaultModelId: GroqModelId = "moonshotai/kimi-k2-instruct-0905"
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Breaking Change: Changing the default model from llama-3.3-70b-versatile to moonshotai/kimi-k2-instruct-0905 will affect all users who rely on the default model. They may experience different behavior, costs, and capabilities. Could we consider:

  1. Keeping the existing default for backward compatibility
  2. Or documenting this breaking change prominently in the changelog/release notes?

contextWindow: 262144,
supportsImages: false,
supportsPromptCache: true,
inputPrice: 0.6,
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Is the significant price reduction intentional? The new model has:

  • Input price: /bin/sh.6 (40% lower than the existing kimi-k2-instruct at .0)
  • Output price: .5 (17% lower than the existing kimi-k2-instruct at .0)

Could you confirm these prices are correct according to Groq's pricing?

supportsPromptCache: true,
inputPrice: 0.6,
outputPrice: 2.5,
cacheReadsPrice: 0.15,
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The cache reads price of /bin/sh.15 represents a 75% discount from the input price (/bin/sh.6), while the existing kimi-k2-instruct model has a 50% discount. Is this aggressive discount structure correct?

inputPrice: 0.6,
outputPrice: 2.5,
cacheReadsPrice: 0.15,
description: "Moonshot AI Kimi K2 Instruct 1T model, 256K context.",
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Minor inconsistency: The description says "256K context" but the contextWindow value is 262144 (which is 256Ki or 256 * 1024). While technically correct, this differs from the pattern used in the existing model which says "128K context" with value 131072. Should we use "256Ki context" or keep it consistent with the existing pattern?

@hannesrudolph hannesrudolph added the Issue/PR - Triage New issue. Needs quick review to confirm validity and assign labels. label Sep 5, 2025
@daniel-lxs daniel-lxs closed this Sep 5, 2025
@github-project-automation github-project-automation bot moved this from New to Done in Roo Code Roadmap Sep 5, 2025
@github-project-automation github-project-automation bot moved this from Triage to Done in Roo Code Roadmap Sep 5, 2025
@daniel-lxs daniel-lxs deleted the groq_new_kimi branch September 5, 2025 20:56
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

enhancement New feature or request Issue/PR - Triage New issue. Needs quick review to confirm validity and assign labels. size:S This PR changes 10-29 lines, ignoring generated files.

Projects

Archived in project

Development

Successfully merging this pull request may close these issues.

4 participants