@@ -62,7 +62,7 @@ type ChatCommand = {
6262 repeatFrequencyPenalty ?: number ,
6363 repeatPresencePenalty ?: number ,
6464 maxTokens : number ,
65- thoughtBudget ?: number ,
65+ reasoningBudget ?: number ,
6666 noHistory : boolean ,
6767 environmentFunctions : boolean ,
6868 tokenPredictionDraftModel ?: string ,
@@ -263,8 +263,8 @@ export const ChatCommand: CommandModule<object, ChatCommand> = {
263263 default : 0 ,
264264 description : "Maximum number of tokens to generate in responses. Set to `0` to disable. Set to `-1` to set to the context size"
265265 } )
266- . option ( "thoughtBudget " , {
267- alias : [ "tb" , "thinkingBudget" , "reasoningBudget " ] ,
266+ . option ( "reasoningBudget " , {
267+ alias : [ "tb" , "thinkingBudget" , "thoughtsBudget " ] ,
268268 type : "number" ,
269269 default : - 1 ,
270270 defaultDescription : "Unlimited" ,
@@ -326,7 +326,7 @@ export const ChatCommand: CommandModule<object, ChatCommand> = {
326326 promptFile, wrapper, noJinja, contextSize, batchSize, flashAttention, swaFullCache,
327327 noTrimWhitespace, grammar, jsonSchemaGrammarFile, threads, temperature, minP, topK,
328328 topP, seed, gpuLayers, repeatPenalty, lastTokensRepeatPenalty, penalizeRepeatingNewLine,
329- repeatFrequencyPenalty, repeatPresencePenalty, maxTokens, thoughtBudget , noHistory,
329+ repeatFrequencyPenalty, repeatPresencePenalty, maxTokens, reasoningBudget , noHistory,
330330 environmentFunctions, tokenPredictionDraftModel, tokenPredictionModelContextSize, debug, meter, timing, noMmap, printTimings
331331 } ) {
332332 try {
@@ -335,7 +335,7 @@ export const ChatCommand: CommandModule<object, ChatCommand> = {
335335 batchSize, flashAttention, swaFullCache, noTrimWhitespace, grammar, jsonSchemaGrammarFile, threads,
336336 temperature, minP, topK, topP, seed,
337337 gpuLayers, lastTokensRepeatPenalty, repeatPenalty, penalizeRepeatingNewLine, repeatFrequencyPenalty, repeatPresencePenalty,
338- maxTokens, thoughtBudget , noHistory, environmentFunctions, tokenPredictionDraftModel, tokenPredictionModelContextSize,
338+ maxTokens, reasoningBudget , noHistory, environmentFunctions, tokenPredictionDraftModel, tokenPredictionModelContextSize,
339339 debug, meter, timing, noMmap, printTimings
340340 } ) ;
341341 } catch ( err ) {
@@ -352,12 +352,12 @@ async function RunChat({
352352 contextSize, batchSize, flashAttention, swaFullCache, noTrimWhitespace, grammar : grammarArg ,
353353 jsonSchemaGrammarFile : jsonSchemaGrammarFilePath ,
354354 threads, temperature, minP, topK, topP, seed, gpuLayers, lastTokensRepeatPenalty, repeatPenalty, penalizeRepeatingNewLine,
355- repeatFrequencyPenalty, repeatPresencePenalty, maxTokens, thoughtBudget , noHistory, environmentFunctions, tokenPredictionDraftModel,
355+ repeatFrequencyPenalty, repeatPresencePenalty, maxTokens, reasoningBudget , noHistory, environmentFunctions, tokenPredictionDraftModel,
356356 tokenPredictionModelContextSize, debug, meter, timing, noMmap, printTimings
357357} : ChatCommand ) {
358358 if ( contextSize === - 1 ) contextSize = undefined ;
359359 if ( gpuLayers === - 1 ) gpuLayers = undefined ;
360- if ( thoughtBudget === - 1 ) thoughtBudget = undefined ;
360+ if ( reasoningBudget === - 1 ) reasoningBudget = undefined ;
361361
362362 const headers = resolveHeaderFlag ( headerArg ) ;
363363 const trimWhitespace = ! noTrimWhitespace ;
@@ -696,7 +696,7 @@ async function RunChat({
696696 signal : abortController . signal ,
697697 stopOnAbortSignal : true ,
698698 budgets : {
699- thoughtTokens : thoughtBudget
699+ thoughtTokens : reasoningBudget
700700 } ,
701701 repeatPenalty : {
702702 penalty : repeatPenalty ,
0 commit comments