Skip to content

Commit 0200d5a

Browse files
prep for PR
1 parent d9595a5 commit 0200d5a

File tree

3 files changed

+167
-183
lines changed

3 files changed

+167
-183
lines changed

src/api/providers/bedrock.ts

Lines changed: 40 additions & 56 deletions
Original file line numberDiff line numberDiff line change
@@ -227,12 +227,12 @@ export class AwsBedrockHandler extends BaseProvider implements SingleCompletionH
227227

228228
// Log the payload for debugging custom ARN issues
229229
if (this.options.awsCustomArn) {
230-
logger.debug("Using custom ARN for Bedrock request", {
231-
ctx: "bedrock",
232-
customArn: this.options.awsCustomArn,
233-
clientRegion: this.client.config.region,
234-
payload: JSON.stringify(payload, null, 2),
235-
})
230+
// logger.debug("Using custom ARN for Bedrock request", {
231+
// ctx: "bedrock",
232+
// customArn: this.options.awsCustomArn,
233+
// clientRegion: this.client.config.region,
234+
// payload: JSON.stringify(payload, null, 2),
235+
// })
236236
}
237237

238238
const command = new ConverseStreamCommand(payload)
@@ -267,15 +267,15 @@ export class AwsBedrockHandler extends BaseProvider implements SingleCompletionH
267267
const cacheReadTokens = usage.cacheReadInputTokens || usage.cacheReadInputTokenCount || 0
268268
const cacheWriteTokens = usage.cacheWriteInputTokens || usage.cacheWriteInputTokenCount || 0
269269

270-
logger.debug("Bedrock usage amounts before yielding", {
271-
ctx: "bedrock",
272-
inputTokens: usage.inputTokens || 0,
273-
outputTokens: usage.outputTokens || 0,
274-
cacheReadTokens,
275-
cacheWriteTokens,
276-
totalTokens: (usage.inputTokens || 0) + (usage.outputTokens || 0),
277-
modelId: modelId,
278-
})
270+
// logger.debug("Bedrock usage amounts before yielding", {
271+
// ctx: "bedrock",
272+
// inputTokens: usage.inputTokens || 0,
273+
// outputTokens: usage.outputTokens || 0,
274+
// cacheReadTokens,
275+
// cacheWriteTokens,
276+
// totalTokens: (usage.inputTokens || 0) + (usage.outputTokens || 0),
277+
// modelId: modelId,
278+
// })
279279

280280
// In test environments, don't include cache tokens to match test expectations
281281
const isTestEnvironment = process.env.NODE_ENV === "test"
@@ -319,15 +319,15 @@ export class AwsBedrockHandler extends BaseProvider implements SingleCompletionH
319319
if (streamEvent?.trace?.promptRouter?.usage) {
320320
const routerUsage = streamEvent.trace.promptRouter.usage
321321

322-
logger.debug("Bedrock prompt router usage amounts before yielding", {
323-
ctx: "bedrock",
324-
inputTokens: routerUsage.inputTokens || 0,
325-
outputTokens: routerUsage.outputTokens || 0,
326-
cacheReadTokens: routerUsage.cacheReadTokens || 0,
327-
cacheWriteTokens: routerUsage.cacheWriteTokens || 0,
328-
totalTokens: (routerUsage.inputTokens || 0) + (routerUsage.outputTokens || 0),
329-
invokedModelId: streamEvent.trace.promptRouter.invokedModelId,
330-
})
322+
// logger.debug("Bedrock prompt router usage amounts before yielding", {
323+
// ctx: "bedrock",
324+
// inputTokens: routerUsage.inputTokens || 0,
325+
// outputTokens: routerUsage.outputTokens || 0,
326+
// cacheReadTokens: routerUsage.cacheReadTokens || 0,
327+
// cacheWriteTokens: routerUsage.cacheWriteTokens || 0,
328+
// totalTokens: (routerUsage.inputTokens || 0) + (routerUsage.outputTokens || 0),
329+
// invokedModelId: streamEvent.trace.promptRouter.invokedModelId,
330+
// })
331331

332332
yield {
333333
type: "usage",
@@ -794,27 +794,12 @@ export class AwsBedrockHandler extends BaseProvider implements SingleCompletionH
794794
messageTemplate: `Request was throttled or rate limited. Please try:
795795
1. Reducing the frequency of requests
796796
2. If using a provisioned model, check its throughput settings
797-
3. Contact AWS support to request a quota increase if needed
798-
799-
Error Details:
800-
{formattedErrorDetails}
801-
802-
Model Information:
803-
- Model ID: {modelId}
804-
- Context window: {contextWindow} tokens`,
797+
3. Contact AWS support to request a quota increase if needed`,
805798
logLevel: "error",
806799
},
807800
TOO_MANY_TOKENS: {
808801
patterns: ["too many tokens"],
809802
messageTemplate: `"Too many tokens" error detected.
810-
811-
Error Details:
812-
{formattedErrorDetails}
813-
814-
Model Information:
815-
- Model ID: {modelId}
816-
- Context window: {contextWindow} tokens
817-
818803
Possible Causes:
819804
1. Input exceeds model's context window limit
820805
2. Rate limiting (too many tokens per minute)
@@ -831,27 +816,13 @@ Suggestions:
831816
},
832817
ABORT: {
833818
patterns: ["aborterror"], // This will match error.name.toLowerCase() for AbortError
834-
messageTemplate: `Request was aborted: The operation timed out or was manually cancelled. Please try again or check your network connection.
835-
836-
Error Details:
837-
{formattedErrorDetails}
838-
839-
Model Information:
840-
- Model ID: {modelId}
841-
- Context window: {contextWindow} tokens`,
819+
messageTemplate: `Request was aborted: The operation timed out or was manually cancelled. Please try again or check your network connection.`,
842820
logLevel: "info",
843821
},
844822
// Default/generic error
845823
GENERIC: {
846824
patterns: [], // Empty patterns array means this is the default
847-
messageTemplate: `
848-
Error Details:
849-
{formattedErrorDetails}
850-
851-
Model Information:
852-
- Model ID: {modelId}
853-
- Context window: {contextWindow} tokens
854-
`,
825+
messageTemplate: `Unknown Error`,
855826
logLevel: "error",
856827
},
857828
}
@@ -959,6 +930,19 @@ Model Information:
959930
): string | Array<{ type: string; text?: string; inputTokens?: number; outputTokens?: number }> {
960931
const isStreamContext = context === "createMessage"
961932

933+
// Check for specific invalid ARN format errors
934+
if (error instanceof Error && error.message.startsWith("INVALID_ARN_FORMAT:")) {
935+
// For completePrompt, return just "Invalid ARN format" without the prefix
936+
if (!isStreamContext) {
937+
return "Invalid ARN format"
938+
}
939+
// For createMessage, return the formatted error
940+
return [
941+
{ type: "text", text: "Error: Invalid ARN format" },
942+
{ type: "usage", inputTokens: 0, outputTokens: 0 },
943+
]
944+
}
945+
962946
// Determine error type
963947
const errorType = this.getErrorType(error)
964948

src/api/transform/cache-strategy/base-strategy.ts

Lines changed: 12 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -149,23 +149,23 @@ export abstract class CacheStrategy {
149149
protected applyCachePoints(messages: Message[], placements: CachePointPlacement[]): Message[] {
150150
const result: Message[] = []
151151

152-
logger.debug("Applying cache points to messages", {
153-
ctx: "cache-strategy",
154-
messageCount: messages.length,
155-
placementCount: placements.length,
156-
placements: placements.map((p) => ({ index: p.index, tokensCovered: p.tokensCovered })),
157-
})
152+
// logger.debug("Applying cache points to messages", {
153+
// ctx: "cache-strategy",
154+
// messageCount: messages.length,
155+
// placementCount: placements.length,
156+
// placements: placements.map((p) => ({ index: p.index, tokensCovered: p.tokensCovered })),
157+
// })
158158

159159
for (let i = 0; i < messages.length; i++) {
160160
const placement = placements.find((p) => p.index === i)
161161

162162
if (placement) {
163-
logger.debug("Adding cache point to message", {
164-
ctx: "cache-strategy",
165-
messageIndex: i,
166-
messageRole: messages[i].role,
167-
tokensCovered: placement.tokensCovered,
168-
})
163+
// logger.debug("Adding cache point to message", {
164+
// ctx: "cache-strategy",
165+
// messageIndex: i,
166+
// messageRole: messages[i].role,
167+
// tokensCovered: placement.tokensCovered,
168+
// })
169169
messages[i].content?.push(this.createCachePoint())
170170
}
171171
result.push(messages[i])

0 commit comments

Comments
 (0)