@@ -1996,6 +1996,7 @@ protected void doTestBedrockRuntimeAi21Jamba() {
19961996 List .of (
19971997 assertAttribute (
19981998 SemanticConventionsConstants .GEN_AI_REQUEST_MODEL , "ai21.jamba-1-5-mini-v1:0" ),
1999+ assertAttribute (SemanticConventionsConstants .GEN_AI_SYSTEM , "aws.bedrock" ),
19992000 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TEMPERATURE , "0.7" ),
20002001 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TOP_P , "0.8" ),
20012002 assertAttribute (SemanticConventionsConstants .GEN_AI_RESPONSE_FINISH_REASONS , "[stop]" ),
@@ -2070,6 +2071,7 @@ protected void doTestBedrockRuntimeAmazonTitan() {
20702071 assertAttribute (
20712072 SemanticConventionsConstants .GEN_AI_REQUEST_MODEL ,
20722073 "amazon.titan-text-premier-v1:0" ),
2074+ assertAttribute (SemanticConventionsConstants .GEN_AI_SYSTEM , "aws.bedrock" ),
20732075 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_MAX_TOKENS , "100" ),
20742076 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TEMPERATURE , "0.7" ),
20752077 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TOP_P , "0.9" ),
@@ -2148,6 +2150,7 @@ protected void doTestBedrockRuntimeAnthropicClaude() {
21482150 assertAttribute (
21492151 SemanticConventionsConstants .GEN_AI_REQUEST_MODEL ,
21502152 "anthropic.claude-3-haiku-20240307-v1:0" ),
2153+ assertAttribute (SemanticConventionsConstants .GEN_AI_SYSTEM , "aws.bedrock" ),
21512154 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_MAX_TOKENS , "512" ),
21522155 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TEMPERATURE , "0.6" ),
21532156 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TOP_P , "0.53" ),
@@ -2225,6 +2228,7 @@ protected void doTestBedrockRuntimeCohereCommandR() {
22252228 List .of (
22262229 assertAttribute (
22272230 SemanticConventionsConstants .GEN_AI_REQUEST_MODEL , "cohere.command-r-v1:0" ),
2231+ assertAttribute (SemanticConventionsConstants .GEN_AI_SYSTEM , "aws.bedrock" ),
22282232 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_MAX_TOKENS , "4096" ),
22292233 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TEMPERATURE , "0.8" ),
22302234 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TOP_P , "0.45" ),
@@ -2302,6 +2306,7 @@ protected void doTestBedrockRuntimeMetaLlama() {
23022306 List .of (
23032307 assertAttribute (
23042308 SemanticConventionsConstants .GEN_AI_REQUEST_MODEL , "meta.llama3-70b-instruct-v1:0" ),
2309+ assertAttribute (SemanticConventionsConstants .GEN_AI_SYSTEM , "aws.bedrock" ),
23052310 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_MAX_TOKENS , "128" ),
23062311 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TEMPERATURE , "0.1" ),
23072312 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TOP_P , "0.9" ),
@@ -2379,11 +2384,12 @@ protected void doTestBedrockRuntimeMistral() {
23792384 assertAttribute (
23802385 SemanticConventionsConstants .GEN_AI_REQUEST_MODEL ,
23812386 "mistral.mistral-large-2402-v1:0" ),
2387+ assertAttribute (SemanticConventionsConstants .GEN_AI_SYSTEM , "aws.bedrock" ),
23822388 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_MAX_TOKENS , "4096" ),
23832389 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TEMPERATURE , "0.75" ),
23842390 assertAttribute (SemanticConventionsConstants .GEN_AI_REQUEST_TOP_P , "0.25" ),
23852391 assertAttribute (SemanticConventionsConstants .GEN_AI_RESPONSE_FINISH_REASONS , "[stop]" ),
2386- assertAttribute (SemanticConventionsConstants .GEN_AI_USAGE_INPUT_TOKENS , "15 " ),
2392+ assertAttribute (SemanticConventionsConstants .GEN_AI_USAGE_INPUT_TOKENS , "16 " ),
23872393 assertAttribute (SemanticConventionsConstants .GEN_AI_USAGE_OUTPUT_TOKENS , "24" )));
23882394 assertMetricClientAttributes (
23892395 metrics ,
0 commit comments