@@ -1997,6 +1997,7 @@ protected void doTestBedrockRuntimeAi21Jamba() {
19971997 List .of (
19981998 assertAttribute (
19991999 SemanticConventionsConstants .GEN_AI_REQUEST_MODEL , "ai21.jamba-1-5-mini-v1:0" ),
2000+ assertAttribute (SemanticConventionsConstants .GEN_AI_SYSTEM , "aws.bedrock" ),
20002001 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TEMPERATURE , "0.7" ),
20012002 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TOP_P , "0.8" ),
20022003 assertAttribute (SemanticConventionsConstants .GEN_AI_RESPONSE_FINISH_REASONS , "[stop]" ),
@@ -2071,6 +2072,7 @@ protected void doTestBedrockRuntimeAmazonTitan() {
20712072 assertAttribute (
20722073 SemanticConventionsConstants .GEN_AI_REQUEST_MODEL ,
20732074 "amazon.titan-text-premier-v1:0" ),
2075+ assertAttribute (SemanticConventionsConstants .GEN_AI_SYSTEM , "aws.bedrock" ),
20742076 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_MAX_TOKENS , "100" ),
20752077 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TEMPERATURE , "0.7" ),
20762078 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TOP_P , "0.9" ),
@@ -2149,6 +2151,7 @@ protected void doTestBedrockRuntimeAnthropicClaude() {
21492151 assertAttribute (
21502152 SemanticConventionsConstants .GEN_AI_REQUEST_MODEL ,
21512153 "anthropic.claude-3-haiku-20240307-v1:0" ),
2154+ assertAttribute (SemanticConventionsConstants .GEN_AI_SYSTEM , "aws.bedrock" ),
21522155 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_MAX_TOKENS , "512" ),
21532156 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TEMPERATURE , "0.6" ),
21542157 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TOP_P , "0.53" ),
@@ -2226,6 +2229,7 @@ protected void doTestBedrockRuntimeCohereCommandR() {
22262229 List .of (
22272230 assertAttribute (
22282231 SemanticConventionsConstants .GEN_AI_REQUEST_MODEL , "cohere.command-r-v1:0" ),
2232+ assertAttribute (SemanticConventionsConstants .GEN_AI_SYSTEM , "aws.bedrock" ),
22292233 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_MAX_TOKENS , "4096" ),
22302234 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TEMPERATURE , "0.8" ),
22312235 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TOP_P , "0.45" ),
@@ -2303,6 +2307,7 @@ protected void doTestBedrockRuntimeMetaLlama() {
23032307 List .of (
23042308 assertAttribute (
23052309 SemanticConventionsConstants .GEN_AI_REQUEST_MODEL , "meta.llama3-70b-instruct-v1:0" ),
2310+ assertAttribute (SemanticConventionsConstants .GEN_AI_SYSTEM , "aws.bedrock" ),
23062311 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_MAX_TOKENS , "128" ),
23072312 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TEMPERATURE , "0.1" ),
23082313 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TOP_P , "0.9" ),
@@ -2380,11 +2385,12 @@ protected void doTestBedrockRuntimeMistral() {
23802385 assertAttribute (
23812386 SemanticConventionsConstants .GEN_AI_REQUEST_MODEL ,
23822387 "mistral.mistral-large-2402-v1:0" ),
2388+ assertAttribute (SemanticConventionsConstants .GEN_AI_SYSTEM , "aws.bedrock" ),
23832389 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_MAX_TOKENS , "4096" ),
23842390 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TEMPERATURE , "0.75" ),
23852391 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TOP_P , "0.25" ),
23862392 assertAttribute (SemanticConventionsConstants .GEN_AI_RESPONSE_FINISH_REASONS , "[stop]" ),
2387- assertAttribute (SemanticConventionsConstants .GEN_AI_USAGE_INPUT_TOKENS , "15 " ),
2393+ assertAttribute (SemanticConventionsConstants .GEN_AI_USAGE_INPUT_TOKENS , "16 " ),
23882394 assertAttribute (SemanticConventionsConstants .GEN_AI_USAGE_OUTPUT_TOKENS , "24" )));
23892395 assertMetricClientAttributes (
23902396 metrics ,
0 commit comments