@@ -227,12 +227,12 @@ export class AwsBedrockHandler extends BaseProvider implements SingleCompletionH
227227
228228 // Log the payload for debugging custom ARN issues
229229 if ( this . options . awsCustomArn ) {
230- logger . debug ( "Using custom ARN for Bedrock request" , {
231- ctx : "bedrock" ,
232- customArn : this . options . awsCustomArn ,
233- clientRegion : this . client . config . region ,
234- payload : JSON . stringify ( payload , null , 2 ) ,
235- } )
230+ // logger.debug("Using custom ARN for Bedrock request", {
231+ // ctx: "bedrock",
232+ // customArn: this.options.awsCustomArn,
233+ // clientRegion: this.client.config.region,
234+ // payload: JSON.stringify(payload, null, 2),
235+ // })
236236 }
237237
238238 const command = new ConverseStreamCommand ( payload )
@@ -267,15 +267,15 @@ export class AwsBedrockHandler extends BaseProvider implements SingleCompletionH
267267 const cacheReadTokens = usage . cacheReadInputTokens || usage . cacheReadInputTokenCount || 0
268268 const cacheWriteTokens = usage . cacheWriteInputTokens || usage . cacheWriteInputTokenCount || 0
269269
270- logger . debug ( "Bedrock usage amounts before yielding" , {
271- ctx : "bedrock" ,
272- inputTokens : usage . inputTokens || 0 ,
273- outputTokens : usage . outputTokens || 0 ,
274- cacheReadTokens,
275- cacheWriteTokens,
276- totalTokens : ( usage . inputTokens || 0 ) + ( usage . outputTokens || 0 ) ,
277- modelId : modelId ,
278- } )
270+ // logger.debug("Bedrock usage amounts before yielding", {
271+ // ctx: "bedrock",
272+ // inputTokens: usage.inputTokens || 0,
273+ // outputTokens: usage.outputTokens || 0,
274+ // cacheReadTokens,
275+ // cacheWriteTokens,
276+ // totalTokens: (usage.inputTokens || 0) + (usage.outputTokens || 0),
277+ // modelId: modelId,
278+ // })
279279
280280 // In test environments, don't include cache tokens to match test expectations
281281 const isTestEnvironment = process . env . NODE_ENV === "test"
@@ -319,15 +319,15 @@ export class AwsBedrockHandler extends BaseProvider implements SingleCompletionH
319319 if ( streamEvent ?. trace ?. promptRouter ?. usage ) {
320320 const routerUsage = streamEvent . trace . promptRouter . usage
321321
322- logger . debug ( "Bedrock prompt router usage amounts before yielding" , {
323- ctx : "bedrock" ,
324- inputTokens : routerUsage . inputTokens || 0 ,
325- outputTokens : routerUsage . outputTokens || 0 ,
326- cacheReadTokens : routerUsage . cacheReadTokens || 0 ,
327- cacheWriteTokens : routerUsage . cacheWriteTokens || 0 ,
328- totalTokens : ( routerUsage . inputTokens || 0 ) + ( routerUsage . outputTokens || 0 ) ,
329- invokedModelId : streamEvent . trace . promptRouter . invokedModelId ,
330- } )
322+ // logger.debug("Bedrock prompt router usage amounts before yielding", {
323+ // ctx: "bedrock",
324+ // inputTokens: routerUsage.inputTokens || 0,
325+ // outputTokens: routerUsage.outputTokens || 0,
326+ // cacheReadTokens: routerUsage.cacheReadTokens || 0,
327+ // cacheWriteTokens: routerUsage.cacheWriteTokens || 0,
328+ // totalTokens: (routerUsage.inputTokens || 0) + (routerUsage.outputTokens || 0),
329+ // invokedModelId: streamEvent.trace.promptRouter.invokedModelId,
330+ // })
331331
332332 yield {
333333 type : "usage" ,
@@ -794,27 +794,12 @@ export class AwsBedrockHandler extends BaseProvider implements SingleCompletionH
794794 messageTemplate : `Request was throttled or rate limited. Please try:
7957951. Reducing the frequency of requests
7967962. If using a provisioned model, check its throughput settings
797- 3. Contact AWS support to request a quota increase if needed
798-
799- Error Details:
800- {formattedErrorDetails}
801-
802- Model Information:
803- - Model ID: {modelId}
804- - Context window: {contextWindow} tokens` ,
797+ 3. Contact AWS support to request a quota increase if needed` ,
805798 logLevel : "error" ,
806799 } ,
807800 TOO_MANY_TOKENS : {
808801 patterns : [ "too many tokens" ] ,
809802 messageTemplate : `"Too many tokens" error detected.
810-
811- Error Details:
812- {formattedErrorDetails}
813-
814- Model Information:
815- - Model ID: {modelId}
816- - Context window: {contextWindow} tokens
817-
818803Possible Causes:
8198041. Input exceeds model's context window limit
8208052. Rate limiting (too many tokens per minute)
@@ -831,27 +816,13 @@ Suggestions:
831816 } ,
832817 ABORT : {
833818 patterns : [ "aborterror" ] , // This will match error.name.toLowerCase() for AbortError
834- messageTemplate : `Request was aborted: The operation timed out or was manually cancelled. Please try again or check your network connection.
835-
836- Error Details:
837- {formattedErrorDetails}
838-
839- Model Information:
840- - Model ID: {modelId}
841- - Context window: {contextWindow} tokens` ,
819+ messageTemplate : `Request was aborted: The operation timed out or was manually cancelled. Please try again or check your network connection.` ,
842820 logLevel : "info" ,
843821 } ,
844822 // Default/generic error
845823 GENERIC : {
846824 patterns : [ ] , // Empty patterns array means this is the default
847- messageTemplate : `
848- Error Details:
849- {formattedErrorDetails}
850-
851- Model Information:
852- - Model ID: {modelId}
853- - Context window: {contextWindow} tokens
854- ` ,
825+ messageTemplate : `Unknown Error` ,
855826 logLevel : "error" ,
856827 } ,
857828 }
@@ -959,6 +930,19 @@ Model Information:
959930 ) : string | Array < { type : string ; text ?: string ; inputTokens ?: number ; outputTokens ?: number } > {
960931 const isStreamContext = context === "createMessage"
961932
933+ // Check for specific invalid ARN format errors
934+ if ( error instanceof Error && error . message . startsWith ( "INVALID_ARN_FORMAT:" ) ) {
935+ // For completePrompt, return just "Invalid ARN format" without the prefix
936+ if ( ! isStreamContext ) {
937+ return "Invalid ARN format"
938+ }
939+ // For createMessage, return the formatted error
940+ return [
941+ { type : "text" , text : "Error: Invalid ARN format" } ,
942+ { type : "usage" , inputTokens : 0 , outputTokens : 0 } ,
943+ ]
944+ }
945+
962946 // Determine error type
963947 const errorType = this . getErrorType ( error )
964948
0 commit comments