|
43 | 43 | _GEN_AI_RESPONSE_FINISH_REASONS: str = "gen_ai.response.finish_reasons"
|
44 | 44 | _GEN_AI_USAGE_INPUT_TOKENS: str = "gen_ai.usage.input_tokens"
|
45 | 45 | _GEN_AI_USAGE_OUTPUT_TOKENS: str = "gen_ai.usage.output_tokens"
|
| 46 | +_GEN_AI_SYSTEM: str = "gen_ai.system" |
46 | 47 |
|
47 | 48 | _AWS_SECRET_ARN: str = "aws.secretsmanager.secret.arn"
|
48 | 49 | _AWS_STATE_MACHINE_ARN: str = "aws.stepfunctions.state_machine.arn"
|
@@ -428,6 +429,7 @@ def test_bedrock_runtime_invoke_model_amazon_titan(self):
|
428 | 429 | cloudformation_primary_identifier="amazon.titan-text-premier-v1:0",
|
429 | 430 | request_specific_attributes={
|
430 | 431 | _GEN_AI_REQUEST_MODEL: "amazon.titan-text-premier-v1:0",
|
| 432 | + _GEN_AI_SYSTEM: "aws.bedrock", |
431 | 433 | _GEN_AI_REQUEST_MAX_TOKENS: 3072,
|
432 | 434 | _GEN_AI_REQUEST_TEMPERATURE: 0.7,
|
433 | 435 | _GEN_AI_REQUEST_TOP_P: 0.9,
|
@@ -455,6 +457,7 @@ def test_bedrock_runtime_invoke_model_anthropic_claude(self):
|
455 | 457 | cloudformation_primary_identifier="anthropic.claude-v2:1",
|
456 | 458 | request_specific_attributes={
|
457 | 459 | _GEN_AI_REQUEST_MODEL: "anthropic.claude-v2:1",
|
| 460 | + _GEN_AI_SYSTEM: "aws.bedrock", |
458 | 461 | _GEN_AI_REQUEST_MAX_TOKENS: 1000,
|
459 | 462 | _GEN_AI_REQUEST_TEMPERATURE: 0.99,
|
460 | 463 | _GEN_AI_REQUEST_TOP_P: 1,
|
@@ -482,6 +485,7 @@ def test_bedrock_runtime_invoke_model_meta_llama(self):
|
482 | 485 | cloudformation_primary_identifier="meta.llama2-13b-chat-v1",
|
483 | 486 | request_specific_attributes={
|
484 | 487 | _GEN_AI_REQUEST_MODEL: "meta.llama2-13b-chat-v1",
|
| 488 | + _GEN_AI_SYSTEM: "aws.bedrock", |
485 | 489 | _GEN_AI_REQUEST_MAX_TOKENS: 512,
|
486 | 490 | _GEN_AI_REQUEST_TEMPERATURE: 0.5,
|
487 | 491 | _GEN_AI_REQUEST_TOP_P: 0.9,
|
@@ -509,6 +513,7 @@ def test_bedrock_runtime_invoke_model_cohere_command(self):
|
509 | 513 | cloudformation_primary_identifier="cohere.command-r-v1:0",
|
510 | 514 | request_specific_attributes={
|
511 | 515 | _GEN_AI_REQUEST_MODEL: "cohere.command-r-v1:0",
|
| 516 | + _GEN_AI_SYSTEM: "aws.bedrock", |
512 | 517 | _GEN_AI_REQUEST_MAX_TOKENS: 512,
|
513 | 518 | _GEN_AI_REQUEST_TEMPERATURE: 0.5,
|
514 | 519 | _GEN_AI_REQUEST_TOP_P: 0.65,
|
@@ -538,6 +543,7 @@ def test_bedrock_runtime_invoke_model_ai21_jamba(self):
|
538 | 543 | cloudformation_primary_identifier="ai21.jamba-1-5-large-v1:0",
|
539 | 544 | request_specific_attributes={
|
540 | 545 | _GEN_AI_REQUEST_MODEL: "ai21.jamba-1-5-large-v1:0",
|
| 546 | + _GEN_AI_SYSTEM: "aws.bedrock", |
541 | 547 | _GEN_AI_REQUEST_MAX_TOKENS: 512,
|
542 | 548 | _GEN_AI_REQUEST_TEMPERATURE: 0.6,
|
543 | 549 | _GEN_AI_REQUEST_TOP_P: 0.8,
|
@@ -565,6 +571,7 @@ def test_bedrock_runtime_invoke_model_mistral(self):
|
565 | 571 | cloudformation_primary_identifier="mistral.mistral-7b-instruct-v0:2",
|
566 | 572 | request_specific_attributes={
|
567 | 573 | _GEN_AI_REQUEST_MODEL: "mistral.mistral-7b-instruct-v0:2",
|
| 574 | + _GEN_AI_SYSTEM: "aws.bedrock", |
568 | 575 | _GEN_AI_REQUEST_MAX_TOKENS: 4096,
|
569 | 576 | _GEN_AI_REQUEST_TEMPERATURE: 0.75,
|
570 | 577 | _GEN_AI_REQUEST_TOP_P: 0.99,
|
|
0 commit comments