Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -1997,6 +1997,7 @@ protected void doTestBedrockRuntimeAi21Jamba() {
List.of(
assertAttribute(
SemanticConventionsConstants.GEN_AI_REQUEST_MODEL, "ai21.jamba-1-5-mini-v1:0"),
assertAttribute(SemanticConventionsConstants.GEN_AI_SYSTEM, "aws.bedrock"),
assertAttribute(SemanticConventionsConstants.GEN_AI_REQUEST_TEMPERATURE, "0.7"),
assertAttribute(SemanticConventionsConstants.GEN_AI_REQUEST_TOP_P, "0.8"),
assertAttribute(SemanticConventionsConstants.GEN_AI_RESPONSE_FINISH_REASONS, "[stop]"),
Expand Down Expand Up @@ -2071,6 +2072,7 @@ protected void doTestBedrockRuntimeAmazonTitan() {
assertAttribute(
SemanticConventionsConstants.GEN_AI_REQUEST_MODEL,
"amazon.titan-text-premier-v1:0"),
assertAttribute(SemanticConventionsConstants.GEN_AI_SYSTEM, "aws.bedrock"),
assertAttribute(SemanticConventionsConstants.GEN_AI_REQUEST_MAX_TOKENS, "100"),
assertAttribute(SemanticConventionsConstants.GEN_AI_REQUEST_TEMPERATURE, "0.7"),
assertAttribute(SemanticConventionsConstants.GEN_AI_REQUEST_TOP_P, "0.9"),
Expand Down Expand Up @@ -2149,6 +2151,7 @@ protected void doTestBedrockRuntimeAnthropicClaude() {
assertAttribute(
SemanticConventionsConstants.GEN_AI_REQUEST_MODEL,
"anthropic.claude-3-haiku-20240307-v1:0"),
assertAttribute(SemanticConventionsConstants.GEN_AI_SYSTEM, "aws.bedrock"),
assertAttribute(SemanticConventionsConstants.GEN_AI_REQUEST_MAX_TOKENS, "512"),
assertAttribute(SemanticConventionsConstants.GEN_AI_REQUEST_TEMPERATURE, "0.6"),
assertAttribute(SemanticConventionsConstants.GEN_AI_REQUEST_TOP_P, "0.53"),
Expand Down Expand Up @@ -2226,6 +2229,7 @@ protected void doTestBedrockRuntimeCohereCommandR() {
List.of(
assertAttribute(
SemanticConventionsConstants.GEN_AI_REQUEST_MODEL, "cohere.command-r-v1:0"),
assertAttribute(SemanticConventionsConstants.GEN_AI_SYSTEM, "aws.bedrock"),
assertAttribute(SemanticConventionsConstants.GEN_AI_REQUEST_MAX_TOKENS, "4096"),
assertAttribute(SemanticConventionsConstants.GEN_AI_REQUEST_TEMPERATURE, "0.8"),
assertAttribute(SemanticConventionsConstants.GEN_AI_REQUEST_TOP_P, "0.45"),
Expand Down Expand Up @@ -2303,6 +2307,7 @@ protected void doTestBedrockRuntimeMetaLlama() {
List.of(
assertAttribute(
SemanticConventionsConstants.GEN_AI_REQUEST_MODEL, "meta.llama3-70b-instruct-v1:0"),
assertAttribute(SemanticConventionsConstants.GEN_AI_SYSTEM, "aws.bedrock"),
assertAttribute(SemanticConventionsConstants.GEN_AI_REQUEST_MAX_TOKENS, "128"),
assertAttribute(SemanticConventionsConstants.GEN_AI_REQUEST_TEMPERATURE, "0.1"),
assertAttribute(SemanticConventionsConstants.GEN_AI_REQUEST_TOP_P, "0.9"),
Expand Down Expand Up @@ -2380,11 +2385,12 @@ protected void doTestBedrockRuntimeMistral() {
assertAttribute(
SemanticConventionsConstants.GEN_AI_REQUEST_MODEL,
"mistral.mistral-large-2402-v1:0"),
assertAttribute(SemanticConventionsConstants.GEN_AI_SYSTEM, "aws.bedrock"),
assertAttribute(SemanticConventionsConstants.GEN_AI_REQUEST_MAX_TOKENS, "4096"),
assertAttribute(SemanticConventionsConstants.GEN_AI_REQUEST_TEMPERATURE, "0.75"),
assertAttribute(SemanticConventionsConstants.GEN_AI_REQUEST_TOP_P, "0.25"),
assertAttribute(SemanticConventionsConstants.GEN_AI_RESPONSE_FINISH_REASONS, "[stop]"),
assertAttribute(SemanticConventionsConstants.GEN_AI_USAGE_INPUT_TOKENS, "15"),
assertAttribute(SemanticConventionsConstants.GEN_AI_USAGE_INPUT_TOKENS, "16"),
assertAttribute(SemanticConventionsConstants.GEN_AI_USAGE_OUTPUT_TOKENS, "24")));
assertMetricClientAttributes(
metrics,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -64,6 +64,7 @@ public class SemanticConventionsConstants {
public static final String AWS_GUARDRAIL_ID = "aws.bedrock.guardrail.id";
public static final String AWS_GUARDRAIL_ARN = "aws.bedrock.guardrail.arn";
public static final String GEN_AI_REQUEST_MODEL = "gen_ai.request.model";
public static final String GEN_AI_SYSTEM = "gen_ai.system";
public static final String GEN_AI_REQUEST_MAX_TOKENS = "gen_ai.request.max_tokens";
public static final String GEN_AI_REQUEST_TEMPERATURE = "gen_ai.request.temperature";
public static final String GEN_AI_REQUEST_TOP_P = "gen_ai.request.top_p";
Expand Down
Loading