-
Notifications
You must be signed in to change notification settings - Fork 2.5k
feat: add GLM-4.6-turbo model to Chutes provider #8514
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Adds the zai-org/GLM-4.6-turbo model with 200K+ token context window and pricing information to the Chutes provider integration.
Updates the GLM-4.6-FP8 model pricing from 0 to correct values: inputPrice: 0.5, outputPrice: 1.75 to match the chutes API pricing information.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I found some issues that need attention. See inline comments.
| description: | ||
| "GLM-4.6-turbo model with 200K+ token context window, optimized for fast inference with enhanced reasoning and tool use capabilities.", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
P3: Description says "200K+" while contextWindow is 202752. For consistency and clarity, consider using the exact number in the description (e.g., "202,752-token context window").
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I found some issues that need attention. See inline comments.
| description: | ||
| "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios.", | ||
| }, | ||
| "zai-org/GLM-4.6-turbo": { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
P2: Add/update tests to cover this new model ID in the Chutes provider. Include assertions that the model appears in the supported IDs and that its ModelInfo (contextWindow, maxTokens, pricing, image/prompt-cache support) matches expectations, e.g., in src/api/providers/tests/chutes.spec.ts.
Adds the zai-org/GLM-4.6-turbo model with 200K+ token context window and pricing information to the Chutes provider integration.
Important
Adds
zai-org/GLM-4.6-turbomodel to Chutes provider with 200K+ token context window and pricing details.zai-org/GLM-4.6-turbotoChutesModelIdinchutes.ts.chutesModelsinchutes.tswithGLM-4.6-turbospecifications: 32768 max tokens, 202752 context window, input price 1.15, output price 3.25.GLM-4.6-turboas optimized for fast inference with enhanced reasoning and tool use capabilities.This description was created by
for 6c83f94. You can customize this summary. It will automatically update as commits are pushed.