@@ -10,6 +10,8 @@ import {
1010import { ApiHandler , SingleCompletionHandler } from "../index"
1111import { ApiStream } from "../transform/stream"
1212
13+ const ANTHROPIC_DEFAULT_TEMPERATURE = 0
14+
1315export class AnthropicHandler implements ApiHandler , SingleCompletionHandler {
1416 private options : ApiHandlerOptions
1517 private client : Anthropic
@@ -44,7 +46,7 @@ export class AnthropicHandler implements ApiHandler, SingleCompletionHandler {
4446 {
4547 model : modelId ,
4648 max_tokens : this . getModel ( ) . info . maxTokens || 8192 ,
47- temperature : this . options . modelTemperature ?? 0 ,
49+ temperature : this . options . modelTemperature ?? ANTHROPIC_DEFAULT_TEMPERATURE ,
4850 system : [ { text : systemPrompt , type : "text" , cache_control : { type : "ephemeral" } } ] , // setting cache breakpoint for system prompt so new tasks can reuse it
4951 messages : messages . map ( ( message , index ) => {
5052 if ( index === lastUserMsgIndex || index === secondLastMsgUserIndex ) {
@@ -96,7 +98,7 @@ export class AnthropicHandler implements ApiHandler, SingleCompletionHandler {
9698 stream = ( await this . client . messages . create ( {
9799 model : modelId ,
98100 max_tokens : this . getModel ( ) . info . maxTokens || 8192 ,
99- temperature : this . options . modelTemperature ?? 0 ,
101+ temperature : this . options . modelTemperature ?? ANTHROPIC_DEFAULT_TEMPERATURE ,
100102 system : [ { text : systemPrompt , type : "text" } ] ,
101103 messages,
102104 // tools,
@@ -179,7 +181,7 @@ export class AnthropicHandler implements ApiHandler, SingleCompletionHandler {
179181 const response = await this . client . messages . create ( {
180182 model : this . getModel ( ) . id ,
181183 max_tokens : this . getModel ( ) . info . maxTokens || 8192 ,
182- temperature : this . options . modelTemperature ?? 0 ,
184+ temperature : this . options . modelTemperature ?? ANTHROPIC_DEFAULT_TEMPERATURE ,
183185 messages : [ { role : "user" , content : prompt } ] ,
184186 stream : false ,
185187 } )
0 commit comments