Replies: 1 comment 2 replies
-
Hey, don't want to clutter the Issues page with suggestions like this but I would very much appreciate some feedback if this is doable/within scope @LostRuins 🙇🏻♂️ |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
The thinking process in DeepSeek-derived models can often be longer than the final result, but the user has no way to anticipate this, which often leads to just an incomplete thinking process with no final answer.
It would be very nice to include an option where the model can "think" as much as needed (or within a separately defined limit) without counting those tokens toward the final
Max Output
. This would allow the reasoning phase to function more independently since it’s not necessarily intended for the user to see or directly use it.From my attempt to somewhat circumvent this I would also like to know: is there currently a way to define (via arguments) the initial Max Output size? Or the effective range available in the ui slider? I know you can manually type values above 512, but a way to set it like you can
--contextsize
would be welcomed as well.Beta Was this translation helpful? Give feedback.
All reactions