-
Notifications
You must be signed in to change notification settings - Fork 2.8k
feat: add GLM-4.6-turbo model to Chutes provider #8516
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: add GLM-4.6-turbo model to Chutes provider #8516
Conversation
|
@raulcorreia7 its free? |
9888efa to
65f4d63
Compare
|
@hannesrudolph , no. I think its a high QUANT/high TPS model. Worth it with the 20$ subscription https://chutes.ai/app/chute/69f319a7-7f25-5cce-ae30-68b30fd5ec5d?tab=stats |
65f4d63 to
5d36ecf
Compare
|
You have $0 for the price |
|
I could open a seperate issue to update all model prices. You can query https://llm.chutes.api/v1/models and extract all the pricing. |
|
Please check again, I force pushed. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I found some issues that need attention. See inline comments.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I found some issues that need attention.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I found some issues that need attention. See inline comments.
40a7c18 to
d42a3d7
Compare
… zai-org/GLM-4.6-turbo model to Chutes. Closes RooCodeInc#8515
d42a3d7 to
1f3ad22
Compare
|
I was struggling with your linter/formatter. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I found some issues that need attention.
Co-authored-by: roomote[bot] <219738659+roomote[bot]@users.noreply.github.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I found some issues that need attention.
|
@hannesrudolph should be complete IMO. do a final review |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I found some issues that need attention.
Co-authored-by: roomote[bot] <219738659+roomote[bot]@users.noreply.github.com>
Co-authored-by: roomote[bot] <219738659+roomote[bot]@users.noreply.github.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I found some issues that need attention.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I found some issues that need attention.
…mperature assertion Co-authored-by: Qwen-Coder <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I found some issues that need attention.
Co-authored-by: roomote[bot] <219738659+roomote[bot]@users.noreply.github.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
No new blocking issues. One minor test enhancement suggestion to improve regression coverage for the new model.
…atch mapping/regression errors beyond temperature
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
One minor test enhancement suggestion to improve regression coverage for the new model.
| const model = chutesModels["zai-org/GLM-4.6-turbo"] | ||
| expect(model.maxTokens).toBe(32768) | ||
| expect(model.contextWindow).toBe(202752) | ||
| expect(model.supportsImages).toBe(false) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
[P3] Test coverage: Add a createMessage param-passthrough test specifically for "zai-org/GLM-4.6-turbo". The suite already verifies DeepSeek R1 and a generic non‑DeepSeek model; asserting this model’s call (model id, max_tokens 32768, temperature 0.5, stream true, stream_options { include_usage: true }) would catch mapping regressions that the generic case might miss.
…ct value validation - Add compile-time assertion ensuring GLM-4.6-turbo is part of ChutesModelId - Update test to use modelId variable for better type safety - Add comprehensive validation of all model properties with exact values - Change description formatting to multi-line style for consistency
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I found some issues that need attention.
Co-authored-by: roomote[bot] <219738659+roomote[bot]@users.noreply.github.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I found some issues that need attention.
| }), | ||
| ) | ||
| // Strengthen test by also asserting the selected model.info matches the static config to catch mapping/regression errors beyond temperature | ||
| expect(model.info).toEqual(expect.objectContaining(chutesModels[testModelId])) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
[P3] Test duplication: the GLM-4.6-turbo config assertions repeat patterns used elsewhere. Consider a small helper to keep assertions consistent and reduce noise across tests (e.g., expectModelInfo(model.info, chutesModels[testModelId])).
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I found some issues that need attention.
| model: modelId, | ||
| max_tokens: modelInfo.maxTokens, // Should be 32768 | ||
| temperature: 0.5, // Default temperature for non-DeepSeek models | ||
| messages: expect.arrayContaining([{ role: "system", content: systemPrompt }]), |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
[P3] Strengthen assertion: also verify the user message is passed through to the OpenAI client to catch regressions in message composition for non-DeepSeek models (in addition to the system prompt). Consider asserting arrayContaining includes both the system and user entries.
Adds the missing zai-org/GLM-4.6-turbo model to the Chutes provider configuration. Closes #8515
Important
Add
zai-org/GLM-4.6-turbomodel to Chutes provider configuration inchutes.ts.zai-org/GLM-4.6-turbotoChutesModelIdinchutes.ts.zai-org/GLM-4.6-turboinchutesModelswith 32768 max tokens, 202752 context window, no image support, no prompt cache, and 1.15 input / 3.25 output price.This description was created by
for 9888efae3b46d5a16fcf89bd80d210dcaa4121a6. You can customize this summary. It will automatically update as commits are pushed.