|
44 | 44 | _GEN_AI_RESPONSE_FINISH_REASONS: str = "gen_ai.response.finish_reasons" |
45 | 45 | _GEN_AI_USAGE_INPUT_TOKENS: str = "gen_ai.usage.input_tokens" |
46 | 46 | _GEN_AI_USAGE_OUTPUT_TOKENS: str = "gen_ai.usage.output_tokens" |
| 47 | +_GEN_AI_SYSTEM: str = "gen_ai.system" |
47 | 48 |
|
48 | 49 | _AWS_SECRET_ARN: str = "aws.secretsmanager.secret.arn" |
49 | 50 | _AWS_STATE_MACHINE_ARN: str = "aws.stepfunctions.state_machine.arn" |
@@ -429,6 +430,7 @@ def test_bedrock_runtime_invoke_model_amazon_titan(self): |
429 | 430 | cloudformation_primary_identifier="amazon.titan-text-premier-v1:0", |
430 | 431 | request_specific_attributes={ |
431 | 432 | _GEN_AI_REQUEST_MODEL: "amazon.titan-text-premier-v1:0", |
| 433 | + _GEN_AI_SYSTEM: "aws.bedrock", |
432 | 434 | _GEN_AI_REQUEST_MAX_TOKENS: 3072, |
433 | 435 | _GEN_AI_REQUEST_TEMPERATURE: 0.7, |
434 | 436 | _GEN_AI_REQUEST_TOP_P: 0.9, |
@@ -483,6 +485,7 @@ def test_bedrock_runtime_invoke_model_anthropic_claude(self): |
483 | 485 | cloudformation_primary_identifier="anthropic.claude-v2:1", |
484 | 486 | request_specific_attributes={ |
485 | 487 | _GEN_AI_REQUEST_MODEL: "anthropic.claude-v2:1", |
| 488 | + _GEN_AI_SYSTEM: "aws.bedrock", |
486 | 489 | _GEN_AI_REQUEST_MAX_TOKENS: 1000, |
487 | 490 | _GEN_AI_REQUEST_TEMPERATURE: 0.99, |
488 | 491 | _GEN_AI_REQUEST_TOP_P: 1, |
@@ -510,6 +513,7 @@ def test_bedrock_runtime_invoke_model_meta_llama(self): |
510 | 513 | cloudformation_primary_identifier="meta.llama2-13b-chat-v1", |
511 | 514 | request_specific_attributes={ |
512 | 515 | _GEN_AI_REQUEST_MODEL: "meta.llama2-13b-chat-v1", |
| 516 | + _GEN_AI_SYSTEM: "aws.bedrock", |
513 | 517 | _GEN_AI_REQUEST_MAX_TOKENS: 512, |
514 | 518 | _GEN_AI_REQUEST_TEMPERATURE: 0.5, |
515 | 519 | _GEN_AI_REQUEST_TOP_P: 0.9, |
@@ -537,6 +541,7 @@ def test_bedrock_runtime_invoke_model_cohere_command(self): |
537 | 541 | cloudformation_primary_identifier="cohere.command-r-v1:0", |
538 | 542 | request_specific_attributes={ |
539 | 543 | _GEN_AI_REQUEST_MODEL: "cohere.command-r-v1:0", |
| 544 | + _GEN_AI_SYSTEM: "aws.bedrock", |
540 | 545 | _GEN_AI_REQUEST_MAX_TOKENS: 512, |
541 | 546 | _GEN_AI_REQUEST_TEMPERATURE: 0.5, |
542 | 547 | _GEN_AI_REQUEST_TOP_P: 0.65, |
@@ -566,6 +571,7 @@ def test_bedrock_runtime_invoke_model_ai21_jamba(self): |
566 | 571 | cloudformation_primary_identifier="ai21.jamba-1-5-large-v1:0", |
567 | 572 | request_specific_attributes={ |
568 | 573 | _GEN_AI_REQUEST_MODEL: "ai21.jamba-1-5-large-v1:0", |
| 574 | + _GEN_AI_SYSTEM: "aws.bedrock", |
569 | 575 | _GEN_AI_REQUEST_MAX_TOKENS: 512, |
570 | 576 | _GEN_AI_REQUEST_TEMPERATURE: 0.6, |
571 | 577 | _GEN_AI_REQUEST_TOP_P: 0.8, |
@@ -593,6 +599,7 @@ def test_bedrock_runtime_invoke_model_mistral(self): |
593 | 599 | cloudformation_primary_identifier="mistral.mistral-7b-instruct-v0:2", |
594 | 600 | request_specific_attributes={ |
595 | 601 | _GEN_AI_REQUEST_MODEL: "mistral.mistral-7b-instruct-v0:2", |
| 602 | + _GEN_AI_SYSTEM: "aws.bedrock", |
596 | 603 | _GEN_AI_REQUEST_MAX_TOKENS: 4096, |
597 | 604 | _GEN_AI_REQUEST_TEMPERATURE: 0.75, |
598 | 605 | _GEN_AI_REQUEST_TOP_P: 0.99, |
|
0 commit comments