@@ -20,7 +20,7 @@ export class OllamaHandler implements ApiHandler, SingleCompletionHandler {
2020
2121 async * createMessage ( systemPrompt : string , messages : Anthropic . Messages . MessageParam [ ] ) : ApiStream {
2222 const modelId = this . getModel ( ) . id
23- const useR1Format = modelId . toLowerCase ( ) . includes ( ' deepseek-r1' )
23+ const useR1Format = modelId . toLowerCase ( ) . includes ( " deepseek-r1" )
2424 const openAiMessages : OpenAI . Chat . ChatCompletionMessageParam [ ] = [
2525 { role : "system" , content : systemPrompt } ,
2626 ...( useR1Format ? convertToR1Format ( messages ) : convertToOpenAiMessages ( messages ) ) ,
@@ -29,7 +29,7 @@ export class OllamaHandler implements ApiHandler, SingleCompletionHandler {
2929 const stream = await this . client . chat . completions . create ( {
3030 model : this . getModel ( ) . id ,
3131 messages : openAiMessages ,
32- temperature : 0 ,
32+ temperature : this . options . modelTemperature ?? 0 ,
3333 stream : true ,
3434 } )
3535 for await ( const chunk of stream ) {
@@ -53,11 +53,13 @@ export class OllamaHandler implements ApiHandler, SingleCompletionHandler {
5353 async completePrompt ( prompt : string ) : Promise < string > {
5454 try {
5555 const modelId = this . getModel ( ) . id
56- const useR1Format = modelId . toLowerCase ( ) . includes ( ' deepseek-r1' )
56+ const useR1Format = modelId . toLowerCase ( ) . includes ( " deepseek-r1" )
5757 const response = await this . client . chat . completions . create ( {
5858 model : this . getModel ( ) . id ,
59- messages : useR1Format ? convertToR1Format ( [ { role : "user" , content : prompt } ] ) : [ { role : "user" , content : prompt } ] ,
60- temperature : 0 ,
59+ messages : useR1Format
60+ ? convertToR1Format ( [ { role : "user" , content : prompt } ] )
61+ : [ { role : "user" , content : prompt } ] ,
62+ temperature : this . options . modelTemperature ?? 0 ,
6163 stream : false ,
6264 } )
6365 return response . choices [ 0 ] ?. message . content || ""
0 commit comments