-
Notifications
You must be signed in to change notification settings - Fork 2.6k
fix: update DeepSeek models context window to 128k #7269
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
- Updated deepseek-chat and deepseek-reasoner models from 64k to 128k context window - Updated corresponding test expectations - Aligns with DeepSeek API documentation at https://api-docs.deepseek.com/quick_start/pricing/ Fixes #7268
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Reviewing my own code because apparently I trust no one, not even myself.
| "deepseek-chat": { | ||
| maxTokens: 8192, | ||
| contextWindow: 64_000, | ||
| contextWindow: 128_000, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The context window update to 128k looks correct and aligns with the DeepSeek API documentation. Consider adding a comment referencing the API docs URL (https://api-docs.deepseek.com/quick_start/pricing/) for future maintainers.
Also, this capability update might warrant a changelog entry so users know about the increased context window.
|
@roomote-agent read https://api-docs.deepseek.com/zh-cn/quick_start/pricing and update this pr. |
|
Hi @NaccOll! I see your request to review the DeepSeek pricing documentation and update this PR accordingly. Let me check the documentation and make the necessary updates. |
|
Hi @NaccOll! I have reviewed the DeepSeek pricing documentation at https://api-docs.deepseek.com/zh-cn/quick_start/pricing and updated the PR accordingly. Changes Made:
Summary from Documentation:
All tests pass and CI checks are green. The PR now accurately reflects the latest DeepSeek API capabilities. |
- deepseek-chat: 4096 (4K default) instead of 8192 (8K max) - deepseek-reasoner: 32768 (32K default) instead of 65536 (64K max) - Updated tests to match new default values - Updated description to clarify default vs max output tokens
daniel-lxs
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
For reference: https://api-docs.deepseek.com/quick_start/pricing/
- deepseek-chat: 8192 (8K max) - deepseek-reasoner: 65536 (64K max) - Updated tests to match maximum values - Updated description to reflect 64K max output
This PR attempts to address Issue #7268. Feedback and guidance are welcome.
Summary
Updated DeepSeek models (
deepseek-chatanddeepseek-reasoner) context window from 64k to 128k tokens to align with the latest DeepSeek API capabilities.Changes
64_000to128_000for both DeepSeek models inpackages/types/src/providers/deepseek.tssrc/api/providers/__tests__/deepseek.spec.tsContext
According to the DeepSeek API documentation (https://api-docs.deepseek.com/quick_start/pricing/), their models now support 128k context window.
Testing
Fixes #7268
Important
Update DeepSeek models' context window to 128k tokens in
deepseek.tsand adjust tests indeepseek.spec.ts.64_000to128_000tokens fordeepseek-chatanddeepseek-reasonerindeepseek.ts.deepseek.spec.ts.This description was created by
for a57187b. You can customize this summary. It will automatically update as commits are pushed.