|
43 | 43 | _GEN_AI_RESPONSE_FINISH_REASONS: str = "gen_ai.response.finish_reasons" |
44 | 44 | _GEN_AI_USAGE_INPUT_TOKENS: str = "gen_ai.usage.input_tokens" |
45 | 45 | _GEN_AI_USAGE_OUTPUT_TOKENS: str = "gen_ai.usage.output_tokens" |
| 46 | +_GEN_AI_SYSTEM: str = "gen_ai.system" |
46 | 47 |
|
47 | 48 | _AWS_SECRET_ARN: str = "aws.secretsmanager.secret.arn" |
48 | 49 | _AWS_STATE_MACHINE_ARN: str = "aws.stepfunctions.state_machine.arn" |
@@ -428,6 +429,7 @@ def test_bedrock_runtime_invoke_model_amazon_titan(self): |
428 | 429 | cloudformation_primary_identifier="amazon.titan-text-premier-v1:0", |
429 | 430 | request_specific_attributes={ |
430 | 431 | _GEN_AI_REQUEST_MODEL: "amazon.titan-text-premier-v1:0", |
| 432 | + _GEN_AI_SYSTEM: "aws.bedrock", |
431 | 433 | _GEN_AI_REQUEST_MAX_TOKENS: 3072, |
432 | 434 | _GEN_AI_REQUEST_TEMPERATURE: 0.7, |
433 | 435 | _GEN_AI_REQUEST_TOP_P: 0.9, |
@@ -455,6 +457,7 @@ def test_bedrock_runtime_invoke_model_anthropic_claude(self): |
455 | 457 | cloudformation_primary_identifier="anthropic.claude-v2:1", |
456 | 458 | request_specific_attributes={ |
457 | 459 | _GEN_AI_REQUEST_MODEL: "anthropic.claude-v2:1", |
| 460 | + _GEN_AI_SYSTEM: "aws.bedrock", |
458 | 461 | _GEN_AI_REQUEST_MAX_TOKENS: 1000, |
459 | 462 | _GEN_AI_REQUEST_TEMPERATURE: 0.99, |
460 | 463 | _GEN_AI_REQUEST_TOP_P: 1, |
@@ -482,6 +485,7 @@ def test_bedrock_runtime_invoke_model_meta_llama(self): |
482 | 485 | cloudformation_primary_identifier="meta.llama2-13b-chat-v1", |
483 | 486 | request_specific_attributes={ |
484 | 487 | _GEN_AI_REQUEST_MODEL: "meta.llama2-13b-chat-v1", |
| 488 | + _GEN_AI_SYSTEM: "aws.bedrock", |
485 | 489 | _GEN_AI_REQUEST_MAX_TOKENS: 512, |
486 | 490 | _GEN_AI_REQUEST_TEMPERATURE: 0.5, |
487 | 491 | _GEN_AI_REQUEST_TOP_P: 0.9, |
@@ -509,6 +513,7 @@ def test_bedrock_runtime_invoke_model_cohere_command(self): |
509 | 513 | cloudformation_primary_identifier="cohere.command-r-v1:0", |
510 | 514 | request_specific_attributes={ |
511 | 515 | _GEN_AI_REQUEST_MODEL: "cohere.command-r-v1:0", |
| 516 | + _GEN_AI_SYSTEM: "aws.bedrock", |
512 | 517 | _GEN_AI_REQUEST_MAX_TOKENS: 512, |
513 | 518 | _GEN_AI_REQUEST_TEMPERATURE: 0.5, |
514 | 519 | _GEN_AI_REQUEST_TOP_P: 0.65, |
@@ -538,6 +543,7 @@ def test_bedrock_runtime_invoke_model_ai21_jamba(self): |
538 | 543 | cloudformation_primary_identifier="ai21.jamba-1-5-large-v1:0", |
539 | 544 | request_specific_attributes={ |
540 | 545 | _GEN_AI_REQUEST_MODEL: "ai21.jamba-1-5-large-v1:0", |
| 546 | + _GEN_AI_SYSTEM: "aws.bedrock", |
541 | 547 | _GEN_AI_REQUEST_MAX_TOKENS: 512, |
542 | 548 | _GEN_AI_REQUEST_TEMPERATURE: 0.6, |
543 | 549 | _GEN_AI_REQUEST_TOP_P: 0.8, |
@@ -565,6 +571,7 @@ def test_bedrock_runtime_invoke_model_mistral(self): |
565 | 571 | cloudformation_primary_identifier="mistral.mistral-7b-instruct-v0:2", |
566 | 572 | request_specific_attributes={ |
567 | 573 | _GEN_AI_REQUEST_MODEL: "mistral.mistral-7b-instruct-v0:2", |
| 574 | + _GEN_AI_SYSTEM: "aws.bedrock", |
568 | 575 | _GEN_AI_REQUEST_MAX_TOKENS: 4096, |
569 | 576 | _GEN_AI_REQUEST_TEMPERATURE: 0.75, |
570 | 577 | _GEN_AI_REQUEST_TOP_P: 0.99, |
|
0 commit comments