@@ -211,9 +211,7 @@ export class BedrockInstrumentation extends InstrumentationBase {
211211 : { } ;
212212
213213 if ( ATTR_GEN_AI_SYSTEM in attributes ) {
214- const modelId = attributes [
215- ATTR_GEN_AI_RESPONSE_MODEL
216- ] as string ;
214+ const modelId = attributes [ ATTR_GEN_AI_RESPONSE_MODEL ] as string ;
217215 const { modelVendor, model } = this . _extractVendorAndModel ( modelId ) ;
218216
219217 span . setAttribute ( ATTR_GEN_AI_RESPONSE_MODEL , model ) ;
@@ -263,14 +261,11 @@ export class BedrockInstrumentation extends InstrumentationBase {
263261 if ( this . _shouldSendPrompts ( ) ) {
264262 // Update local value with attribute value that was set by _setResponseAttributes
265263 streamedContent +=
266- responseAttributes [
267- `${ ATTR_GEN_AI_COMPLETION } .0.content`
268- ] ;
264+ responseAttributes [ `${ ATTR_GEN_AI_COMPLETION } .0.content` ] ;
269265 // re-assign the new value to responseAttributes
270266 responseAttributes = {
271267 ...responseAttributes ,
272- [ `${ ATTR_GEN_AI_COMPLETION } .0.content` ] :
273- streamedContent ,
268+ [ `${ ATTR_GEN_AI_COMPLETION } .0.content` ] : streamedContent ,
274269 } ;
275270 }
276271
@@ -310,10 +305,8 @@ export class BedrockInstrumentation extends InstrumentationBase {
310305 case "ai21" : {
311306 return {
312307 [ ATTR_GEN_AI_REQUEST_TOP_P ] : requestBody [ "topP" ] ,
313- [ ATTR_GEN_AI_REQUEST_TEMPERATURE ] :
314- requestBody [ "temperature" ] ,
315- [ ATTR_GEN_AI_REQUEST_MAX_TOKENS ] :
316- requestBody [ "maxTokens" ] ,
308+ [ ATTR_GEN_AI_REQUEST_TEMPERATURE ] : requestBody [ "temperature" ] ,
309+ [ ATTR_GEN_AI_REQUEST_MAX_TOKENS ] : requestBody [ "maxTokens" ] ,
317310 [ SpanAttributes . LLM_PRESENCE_PENALTY ] :
318311 requestBody [ "presencePenalty" ] [ "scale" ] ,
319312 [ SpanAttributes . LLM_FREQUENCY_PENALTY ] :
@@ -323,8 +316,7 @@ export class BedrockInstrumentation extends InstrumentationBase {
323316 ...( this . _shouldSendPrompts ( )
324317 ? {
325318 [ `${ ATTR_GEN_AI_PROMPT } .0.role` ] : "user" ,
326- [ `${ ATTR_GEN_AI_PROMPT } .0.content` ] :
327- requestBody [ "prompt" ] ,
319+ [ `${ ATTR_GEN_AI_PROMPT } .0.content` ] : requestBody [ "prompt" ] ,
328320 }
329321 : { } ) ,
330322 } ;
@@ -342,8 +334,7 @@ export class BedrockInstrumentation extends InstrumentationBase {
342334 ...( this . _shouldSendPrompts ( )
343335 ? {
344336 [ `${ ATTR_GEN_AI_PROMPT } .0.role` ] : "user" ,
345- [ `${ ATTR_GEN_AI_PROMPT } .0.content` ] :
346- requestBody [ "inputText" ] ,
337+ [ `${ ATTR_GEN_AI_PROMPT } .0.content` ] : requestBody [ "inputText" ] ,
347338 }
348339 : { } ) ,
349340 } ;
@@ -352,8 +343,7 @@ export class BedrockInstrumentation extends InstrumentationBase {
352343 const baseAttributes = {
353344 [ ATTR_GEN_AI_REQUEST_TOP_P ] : requestBody [ "top_p" ] ,
354345 [ SpanAttributes . LLM_TOP_K ] : requestBody [ "top_k" ] ,
355- [ ATTR_GEN_AI_REQUEST_TEMPERATURE ] :
356- requestBody [ "temperature" ] ,
346+ [ ATTR_GEN_AI_REQUEST_TEMPERATURE ] : requestBody [ "temperature" ] ,
357347 [ ATTR_GEN_AI_REQUEST_MAX_TOKENS ] :
358348 requestBody [ "max_tokens_to_sample" ] || requestBody [ "max_tokens" ] ,
359349 } ;
@@ -366,12 +356,9 @@ export class BedrockInstrumentation extends InstrumentationBase {
366356 if ( requestBody [ "messages" ] ) {
367357 const promptAttributes : Record < string , any > = { } ;
368358 requestBody [ "messages" ] . forEach ( ( message : any , index : number ) => {
369- promptAttributes [
370- `${ ATTR_GEN_AI_PROMPT } .${ index } .role`
371- ] = message . role ;
372- promptAttributes [
373- `${ ATTR_GEN_AI_PROMPT } .${ index } .content`
374- ] =
359+ promptAttributes [ `${ ATTR_GEN_AI_PROMPT } .${ index } .role` ] =
360+ message . role ;
361+ promptAttributes [ `${ ATTR_GEN_AI_PROMPT } .${ index } .content` ] =
375362 typeof message . content === "string"
376363 ? message . content
377364 : JSON . stringify ( message . content ) ;
@@ -384,9 +371,7 @@ export class BedrockInstrumentation extends InstrumentationBase {
384371 return {
385372 ...baseAttributes ,
386373 [ `${ ATTR_GEN_AI_PROMPT } .0.role` ] : "user" ,
387- [ `${ ATTR_GEN_AI_PROMPT } .0.content` ] : requestBody [
388- "prompt"
389- ]
374+ [ `${ ATTR_GEN_AI_PROMPT } .0.content` ] : requestBody [ "prompt" ]
390375 // The format is removing when we are setting span attribute
391376 . replace ( "\n\nHuman:" , "" )
392377 . replace ( "\n\nAssistant:" , "" ) ,
@@ -399,10 +384,8 @@ export class BedrockInstrumentation extends InstrumentationBase {
399384 return {
400385 [ ATTR_GEN_AI_REQUEST_TOP_P ] : requestBody [ "p" ] ,
401386 [ SpanAttributes . LLM_TOP_K ] : requestBody [ "k" ] ,
402- [ ATTR_GEN_AI_REQUEST_TEMPERATURE ] :
403- requestBody [ "temperature" ] ,
404- [ ATTR_GEN_AI_REQUEST_MAX_TOKENS ] :
405- requestBody [ "max_tokens" ] ,
387+ [ ATTR_GEN_AI_REQUEST_TEMPERATURE ] : requestBody [ "temperature" ] ,
388+ [ ATTR_GEN_AI_REQUEST_MAX_TOKENS ] : requestBody [ "max_tokens" ] ,
406389
407390 // Prompt & Role
408391 ...( this . _shouldSendPrompts ( )
@@ -417,17 +400,14 @@ export class BedrockInstrumentation extends InstrumentationBase {
417400 case "meta" : {
418401 return {
419402 [ ATTR_GEN_AI_REQUEST_TOP_P ] : requestBody [ "top_p" ] ,
420- [ ATTR_GEN_AI_REQUEST_TEMPERATURE ] :
421- requestBody [ "temperature" ] ,
422- [ ATTR_GEN_AI_REQUEST_MAX_TOKENS ] :
423- requestBody [ "max_gen_len" ] ,
403+ [ ATTR_GEN_AI_REQUEST_TEMPERATURE ] : requestBody [ "temperature" ] ,
404+ [ ATTR_GEN_AI_REQUEST_MAX_TOKENS ] : requestBody [ "max_gen_len" ] ,
424405
425406 // Prompt & Role
426407 ...( this . _shouldSendPrompts ( )
427408 ? {
428409 [ `${ ATTR_GEN_AI_PROMPT } .0.role` ] : "user" ,
429- [ `${ ATTR_GEN_AI_PROMPT } .0.content` ] :
430- requestBody [ "prompt" ] ,
410+ [ `${ ATTR_GEN_AI_PROMPT } .0.content` ] : requestBody [ "prompt" ] ,
431411 }
432412 : { } ) ,
433413 } ;
@@ -462,8 +442,7 @@ export class BedrockInstrumentation extends InstrumentationBase {
462442 ? response [ "completionReason" ]
463443 : response [ "results" ] [ 0 ] [ "completionReason" ] ,
464444 [ `${ ATTR_GEN_AI_COMPLETION } .0.role` ] : "assistant" ,
465- [ ATTR_GEN_AI_USAGE_PROMPT_TOKENS ] :
466- response [ "inputTextTokenCount" ] ,
445+ [ ATTR_GEN_AI_USAGE_PROMPT_TOKENS ] : response [ "inputTextTokenCount" ] ,
467446 [ ATTR_GEN_AI_USAGE_COMPLETION_TOKENS ] : isStream
468447 ? response [ "totalOutputTextTokenCount" ]
469448 : response [ "results" ] [ 0 ] [ "tokenCount" ] ,
@@ -507,8 +486,7 @@ export class BedrockInstrumentation extends InstrumentationBase {
507486 if ( response [ "completion" ] ) {
508487 return {
509488 ...baseAttributes ,
510- [ `${ ATTR_GEN_AI_COMPLETION } .0.content` ] :
511- response [ "completion" ] ,
489+ [ `${ ATTR_GEN_AI_COMPLETION } .0.content` ] : response [ "completion" ] ,
512490 } ;
513491 }
514492
@@ -532,10 +510,8 @@ export class BedrockInstrumentation extends InstrumentationBase {
532510 const billedUnits = response [ "meta" ] [ "billed_units" ] ;
533511 return {
534512 ...baseAttributes ,
535- [ ATTR_GEN_AI_USAGE_PROMPT_TOKENS ] :
536- billedUnits [ "input_tokens" ] ,
537- [ ATTR_GEN_AI_USAGE_COMPLETION_TOKENS ] :
538- billedUnits [ "output_tokens" ] ,
513+ [ ATTR_GEN_AI_USAGE_PROMPT_TOKENS ] : billedUnits [ "input_tokens" ] ,
514+ [ ATTR_GEN_AI_USAGE_COMPLETION_TOKENS ] : billedUnits [ "output_tokens" ] ,
539515 [ SpanAttributes . LLM_USAGE_TOTAL_TOKENS ] :
540516 ( billedUnits [ "input_tokens" ] || 0 ) +
541517 ( billedUnits [ "output_tokens" ] || 0 ) ,
@@ -549,16 +525,14 @@ export class BedrockInstrumentation extends InstrumentationBase {
549525 [ `${ ATTR_GEN_AI_COMPLETION } .0.finish_reason` ] :
550526 response [ "stop_reason" ] ,
551527 [ `${ ATTR_GEN_AI_COMPLETION } .0.role` ] : "assistant" ,
552- [ ATTR_GEN_AI_USAGE_PROMPT_TOKENS ] :
553- response [ "prompt_token_count" ] ,
528+ [ ATTR_GEN_AI_USAGE_PROMPT_TOKENS ] : response [ "prompt_token_count" ] ,
554529 [ ATTR_GEN_AI_USAGE_COMPLETION_TOKENS ] :
555530 response [ "generation_token_count" ] ,
556531 [ SpanAttributes . LLM_USAGE_TOTAL_TOKENS ] :
557532 response [ "prompt_token_count" ] + response [ "generation_token_count" ] ,
558533 ...( this . _shouldSendPrompts ( )
559534 ? {
560- [ `${ ATTR_GEN_AI_COMPLETION } .0.content` ] :
561- response [ "generation" ] ,
535+ [ `${ ATTR_GEN_AI_COMPLETION } .0.content` ] : response [ "generation" ] ,
562536 }
563537 : { } ) ,
564538 } ;
0 commit comments