Skip to content

Commit 6a39009

Browse files
Add investigation guide for Amazon Bedrock Rules (#4247)
* Add investigation guide for Amazon Bedrock Rules * updated date * review comments * review comments --------- Co-authored-by: Terrance DeJesus <[email protected]>
1 parent 1cc160f commit 6a39009

6 files changed

+213
-7
lines changed

rules/integrations/aws_bedrock/aws_bedrock_guardrails_multiple_violations_by_single_user.toml

Lines changed: 37 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
[metadata]
22
creation_date = "2024/05/02"
33
maturity = "production"
4-
updated_date = "2024/10/09"
4+
updated_date = "2024/11/05"
55
min_stack_comments = "ES|QL rule type is still in technical preview as of 8.13, however this rule was tested successfully; integration in tech preview"
66
min_stack_version = "8.13.0"
77

@@ -26,6 +26,42 @@ references = [
2626
]
2727
risk_score = 47
2828
rule_id = "0cd2f3e6-41da-40e6-b28b-466f688f00a6"
29+
note = """## Triage and analysis
30+
31+
### Investigating Amazon Bedrock Guardrail Multiple Policy Violations by a Single User Over a Session.
32+
33+
Amazon Bedrock Guardrail is a set of features within Amazon Bedrock designed to help businesses apply robust safety and privacy controls to their generative AI applications.
34+
35+
It enables users to set guidelines and filters that manage content quality, relevancy, and adherence to responsible AI practices.
36+
37+
Through Guardrail, organizations can define "denied topics" to prevent the model from generating content on specific, undesired subjects,
38+
and they can establish thresholds for harmful content categories, including hate speech, violence, or offensive language.
39+
40+
#### Possible investigation steps
41+
42+
- Identify the user account that caused multiple policy violations over a session and whether it should perform this kind of action.
43+
- Investigate the user activity that might indicate a potential brute force attack.
44+
- Investigate other alerts associated with the user account during the past 48 hours.
45+
- Consider the time of day. If the user is a human (not a program or script), did the activity take place during a normal time of day?
46+
- Examine the account's prompts and responses in the last 24 hours.
47+
- If you suspect the account has been compromised, scope potentially compromised assets by tracking Amazon Bedrock model access, prompts generated, and responses to the prompts by the account in the last 24 hours.
48+
49+
### False positive analysis
50+
51+
- Verify the user account that caused multiple policy violations by a single user over session, is not testing any new model deployments or updated compliance policies in Amazon Bedrock guardrails.
52+
53+
### Response and remediation
54+
55+
- Initiate the incident response process based on the outcome of the triage.
56+
- Disable or limit the account during the investigation and response.
57+
- Identify the possible impact of the incident and prioritize accordingly; the following actions can help you gain context:
58+
- Identify the account role in the cloud environment.
59+
- Identify if the attacker is moving laterally and compromising other Amazon Bedrock Services.
60+
- Identify any regulatory or legal ramifications related to this activity.
61+
- Review the permissions assigned to the implicated user group or role behind these requests to ensure they are authorized and expected to access bedrock and ensure that the least privilege principle is being followed.
62+
- Determine the initial vector abused by the attacker and take action to prevent reinfection via the same vector.
63+
- Using the incident response data, update logging and audit policies to improve the mean time to detect (MTTD) and the mean time to respond (MTTR).
64+
"""
2965
setup = """## Setup
3066
3167
This rule requires that guardrails are configured in AWS Bedrock. For more information, see the AWS Bedrock documentation:

rules/integrations/aws_bedrock/aws_bedrock_guardrails_multiple_violations_in_single_request.toml

Lines changed: 37 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
[metadata]
22
creation_date = "2024/05/02"
33
maturity = "production"
4-
updated_date = "2024/10/09"
4+
updated_date = "2024/11/05"
55
min_stack_comments = "ES|QL rule type is still in technical preview as of 8.13, however this rule was tested successfully; integration in tech preview"
66
min_stack_version = "8.13.0"
77

@@ -26,6 +26,42 @@ references = [
2626
]
2727
risk_score = 21
2828
rule_id = "f4c2515a-18bb-47ce-a768-1dc4e7b0fe6c"
29+
note = """## Triage and analysis
30+
31+
### Investigating Amazon Bedrock Guardrail Multiple Policy Violations Within a Single Blocked Request.
32+
33+
Amazon Bedrock Guardrail is a set of features within Amazon Bedrock designed to help businesses apply robust safety and privacy controls to their generative AI applications.
34+
35+
It enables users to set guidelines and filters that manage content quality, relevancy, and adherence to responsible AI practices.
36+
37+
Through Guardrail, organizations can define "denied topics" to prevent the model from generating content on specific, undesired subjects,
38+
and they can establish thresholds for harmful content categories, including hate speech, violence, or offensive language.
39+
40+
#### Possible investigation steps
41+
42+
- Identify the user account and the user request that caused multiple policy violations and whether it should perform this kind of action.
43+
- Investigate the user activity that might indicate a potential brute force attack.
44+
- Investigate other alerts associated with the user account during the past 48 hours.
45+
- Consider the time of day. If the user is a human (not a program or script), did the activity take place during a normal time of day?
46+
- Examine the account's prompts and responses in the last 24 hours.
47+
- If you suspect the account has been compromised, scope potentially compromised assets by tracking Amazon Bedrock model access, prompts generated, and responses to the prompts by the account in the last 24 hours.
48+
49+
### False positive analysis
50+
51+
- Verify the user account that caused multiple policy violations, is not testing any new model deployments or updated compliance policies in Amazon Bedrock guardrails.
52+
53+
### Response and remediation
54+
55+
- Initiate the incident response process based on the outcome of the triage.
56+
- Disable or limit the account during the investigation and response.
57+
- Identify the possible impact of the incident and prioritize accordingly; the following actions can help you gain context:
58+
- Identify the account role in the cloud environment.
59+
- Identify if the attacker is moving laterally and compromising other Amazon Bedrock Services.
60+
- Identify any regulatory or legal ramifications related to this activity.
61+
- Review the permissions assigned to the implicated user group or role behind these requests to ensure they are authorized and expected to access bedrock and ensure that the least privilege principle is being followed.
62+
- Determine the initial vector abused by the attacker and take action to prevent reinfection via the same vector.
63+
- Using the incident response data, update logging and audit policies to improve the mean time to detect (MTTD) and the mean time to respond (MTTR).
64+
"""
2965
setup = """## Setup
3066
3167
This rule requires that guardrails are configured in AWS Bedrock. For more information, see the AWS Bedrock documentation:

rules/integrations/aws_bedrock/aws_bedrock_high_confidence_misconduct_blocks_detected.toml

Lines changed: 36 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
[metadata]
22
creation_date = "2024/05/05"
33
maturity = "production"
4-
updated_date = "2024/10/23"
4+
updated_date = "2024/11/05"
55
min_stack_comments = "ES|QL rule type is still in technical preview as of 8.13, however this rule was tested successfully; integration in tech preview"
66
min_stack_version = "8.13.0"
77

@@ -25,6 +25,41 @@ references = [
2525
]
2626
risk_score = 73
2727
rule_id = "4f855297-c8e0-4097-9d97-d653f7e471c4"
28+
note = """## Triage and analysis
29+
30+
### Investigating Amazon Bedrock Guardrail High Confidence Misconduct Blocks.
31+
32+
Amazon Bedrock Guardrail is a set of features within Amazon Bedrock designed to help businesses apply robust safety and privacy controls to their generative AI applications.
33+
34+
It enables users to set guidelines and filters that manage content quality, relevancy, and adherence to responsible AI practices.
35+
36+
Through Guardrail, organizations can define "denied topics" to prevent the model from generating content on specific, undesired subjects,
37+
and they can establish thresholds for harmful content categories, including hate speech, violence, or offensive language.
38+
39+
#### Possible investigation steps
40+
41+
- Identify the user account that queried denied topics and whether it should perform this kind of action.
42+
- Investigate other alerts associated with the user account during the past 48 hours.
43+
- Consider the time of day. If the user is a human (not a program or script), did the activity take place during a normal time of day?
44+
- Examine the account's prompts and responses in the last 24 hours.
45+
- If you suspect the account has been compromised, scope potentially compromised assets by tracking Amazon Bedrock model access, prompts generated, and responses to the prompts by the account in the last 24 hours.
46+
47+
### False positive analysis
48+
49+
- Verify the user account that queried denied topics, is not testing any new model deployments or updated compliance policies in Amazon Bedrock guardrails.
50+
51+
### Response and remediation
52+
53+
- Initiate the incident response process based on the outcome of the triage.
54+
- Disable or limit the account during the investigation and response.
55+
- Identify the possible impact of the incident and prioritize accordingly; the following actions can help you gain context:
56+
- Identify the account role in the cloud environment.
57+
- Identify if the attacker is moving laterally and compromising other Amazon Bedrock Services.
58+
- Identify any regulatory or legal ramifications related to this activity.
59+
- Review the permissions assigned to the implicated user group or role behind these requests to ensure they are authorized and expected to access bedrock and ensure that the least privilege principle is being followed.
60+
- Determine the initial vector abused by the attacker and take action to prevent reinfection via the same vector.
61+
- Using the incident response data, update logging and audit policies to improve the mean time to detect (MTTD) and the mean time to respond (MTTR).
62+
"""
2863
setup = """## Setup
2964
3065
This rule requires that guardrails are configured in AWS Bedrock. For more information, see the AWS Bedrock documentation:

rules/integrations/aws_bedrock/aws_bedrock_high_resource_consumption_detection.toml

Lines changed: 35 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
[metadata]
22
creation_date = "2024/05/04"
33
maturity = "production"
4-
updated_date = "2024/10/09"
4+
updated_date = "2024/11/05"
55
min_stack_comments = "ES|QL rule type is still in technical preview as of 8.13, however this rule was tested successfully; integration in tech preview"
66
min_stack_version = "8.13.0"
77

@@ -25,6 +25,40 @@ references = [
2525
]
2626
risk_score = 47
2727
rule_id = "b1773d05-f349-45fb-9850-287b8f92f02d"
28+
note = """## Triage and analysis
29+
30+
### Investigating Amazon Bedrock Models High Token Count and Large Response Sizes.
31+
32+
Amazon Bedrock is AWS’s managed service that enables developers to build and scale generative AI applications using large foundation models (FMs) from top providers.
33+
34+
Bedrock offers a variety of pretrained models from Amazon (such as the Titan series), as well as models from providers like Anthropic, Meta, Cohere, and AI21 Labs.
35+
36+
#### Possible investigation steps
37+
38+
- Identify the user account that used high prompt token counts and whether it should perform this kind of action.
39+
- Investigate large response sizes and the number of requests made by the user account.
40+
- Investigate other alerts associated with the user account during the past 48 hours.
41+
- Consider the time of day. If the user is a human (not a program or script), did the activity take place during a normal time of day?
42+
- Examine the account's prompts and responses in the last 24 hours.
43+
- If you suspect the account has been compromised, scope potentially compromised assets by tracking Amazon Bedrock model access, prompts generated, and responses to the prompts by the account in the last 24 hours.
44+
45+
### False positive analysis
46+
47+
- Verify the user account that used high prompt and large response sizes, has a business justification for the heavy usage of the system.
48+
49+
### Response and remediation
50+
51+
- Initiate the incident response process based on the outcome of the triage.
52+
- Disable or limit the account during the investigation and response.
53+
- Identify the possible impact of the incident and prioritize accordingly; the following actions can help you gain context:
54+
- Identify the account role in the cloud environment.
55+
- Identify if the attacker is moving laterally and compromising other Amazon Bedrock Services.
56+
- Identify any regulatory or legal ramifications related to this activity.
57+
- Identify potential resource exhaustion and impact on billing.
58+
- Review the permissions assigned to the implicated user group or role behind these requests to ensure they are authorized and expected to access bedrock and ensure that the least privilege principle is being followed.
59+
- Determine the initial vector abused by the attacker and take action to prevent reinfection via the same vector.
60+
- Using the incident response data, update logging and audit policies to improve the mean time to detect (MTTD) and the mean time to respond (MTTR).
61+
"""
2862
setup = """## Setup
2963
3064
This rule requires that guardrails are configured in AWS Bedrock. For more information, see the AWS Bedrock documentation:

rules/integrations/aws_bedrock/aws_bedrock_multiple_attempts_to_use_denied_models_by_user.toml

Lines changed: 33 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
[metadata]
22
creation_date = "2024/05/02"
33
maturity = "production"
4-
updated_date = "2024/10/09"
4+
updated_date = "2024/11/05"
55
min_stack_comments = "ES|QL rule type is still in technical preview as of 8.13, however this rule was tested successfully; integration in tech preview"
66
min_stack_version = "8.13.0"
77

@@ -26,6 +26,38 @@ references = [
2626
]
2727
risk_score = 73
2828
rule_id = "17261da3-a6d0-463c-aac8-ea1718afcd20"
29+
note = """## Triage and analysis
30+
31+
### Investigating Attempt to use Denied Amazon Bedrock Models.
32+
33+
Amazon Bedrock is AWS’s managed service that enables developers to build and scale generative AI applications using large foundation models (FMs) from top providers.
34+
35+
Bedrock offers a variety of pretrained models from Amazon (such as the Titan series), as well as models from providers like Anthropic, Meta, Cohere, and AI21 Labs.
36+
37+
#### Possible investigation steps
38+
39+
- Identify the user account that attempted to use denied models.
40+
- Investigate other alerts associated with the user account during the past 48 hours.
41+
- Consider the time of day. If the user is a human (not a program or script), did the activity take place during a normal time of day?
42+
- Examine the account's attempts to access Amazon Bedrock models in the last 24 hours.
43+
- If you suspect the account has been compromised, scope potentially compromised assets by tracking Amazon Bedrock model access, prompts generated, and responses to the prompts by the account in the last 24 hours.
44+
45+
### False positive analysis
46+
47+
- Verify the user account that attempted to use denied models, is a legitimate misunderstanding by users or overly strict policies.
48+
49+
### Response and remediation
50+
51+
- Initiate the incident response process based on the outcome of the triage.
52+
- Disable or limit the account during the investigation and response.
53+
- Identify the possible impact of the incident and prioritize accordingly; the following actions can help you gain context:
54+
- Identify the account role in the cloud environment.
55+
- Identify if the attacker is moving laterally and compromising other Amazon Bedrock Services.
56+
- Identify any regulatory or legal ramifications related to this activity.
57+
- Review the permissions assigned to the implicated user group or role behind these requests to ensure they are authorized and expected to access bedrock and ensure that the least privilege principle is being followed.
58+
- Determine the initial vector abused by the attacker and take action to prevent reinfection via the same vector.
59+
- Using the incident response data, update logging and audit policies to improve the mean time to detect (MTTD) and the mean time to respond (MTTR).
60+
"""
2961
setup = """## Setup
3062
3163
This rule requires that guardrails are configured in AWS Bedrock. For more information, see the AWS Bedrock documentation:

rules/integrations/aws_bedrock/aws_bedrock_multiple_validation_exception_errors_by_single_user.toml

Lines changed: 35 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
creation_date = "2024/09/11"
33
integration = ["aws_bedrock"]
44
maturity = "production"
5-
updated_date = "2024/10/09"
5+
updated_date = "2024/11/05"
66
min_stack_comments = "ES|QL rule type is still in technical preview as of 8.13, however this rule was tested successfully; integration in tech preview"
77
min_stack_version = "8.13.0"
88

@@ -15,7 +15,7 @@ These errors also occur when you use an inference parameter for one model with a
1515
This could indicate attempts to bypass limitations of other approved models, or to force an impact on the environment by incurring
1616
exhorbitant costs.
1717
"""
18-
false_positives = ["Legitimate misunderstanding by users or overly strict policies"]
18+
false_positives = ["Legitimate misunderstanding by users on accessing the bedrock models."]
1919
from = "now-60m"
2020
interval = "10m"
2121
language = "esql"
@@ -29,6 +29,39 @@ references = [
2929
]
3030
risk_score = 73
3131
rule_id = "725a048a-88c5-4fc7-8677-a44fc0031822"
32+
note = """## Triage and analysis
33+
34+
### Investigating Amazon Bedrock Model Validation Exception Errors.
35+
36+
Amazon Bedrock is AWS’s managed service that enables developers to build and scale generative AI applications using large foundation models (FMs) from top providers.
37+
38+
Bedrock offers a variety of pretrained models from Amazon (such as the Titan series), as well as models from providers like Anthropic, Meta, Cohere, and AI21 Labs.
39+
40+
#### Possible investigation steps
41+
42+
- Identify the user account that caused validation errors in accessing the Amazon Bedrock models.
43+
- Investigate other alerts associated with the user account during the past 48 hours.
44+
- Consider the time of day. If the user is a human (not a program or script), did the activity take place during a normal time of day?
45+
- Examine the account's attempts to access Amazon Bedrock models in the last 24 hours.
46+
- If you suspect the account has been compromised, scope potentially compromised assets by tracking Amazon Bedrock model access, prompts generated, and responses to the prompts by the account in the last 24 hours.
47+
48+
### False positive analysis
49+
50+
- Verify the user account that that caused validation errors is a legitimate misunderstanding by users on accessing the bedrock models.
51+
52+
### Response and remediation
53+
54+
- Initiate the incident response process based on the outcome of the triage.
55+
- Disable or limit the account during the investigation and response.
56+
- Identify the possible impact of the incident and prioritize accordingly; the following actions can help you gain context:
57+
- Identify the account role in the cloud environment.
58+
- Identify if the attacker is moving laterally and compromising other Amazon Bedrock Services.
59+
- Identify any regulatory or legal ramifications related to this activity.
60+
- Identify if any implication to resource billing.
61+
- Review the permissions assigned to the implicated user group or role behind these requests to ensure they are authorized and expected to access bedrock and ensure that the least privilege principle is being followed.
62+
- Determine the initial vector abused by the attacker and take action to prevent reinfection via the same vector.
63+
- Using the incident response data, update logging and audit policies to improve the mean time to detect (MTTD) and the mean time to respond (MTTR).
64+
"""
3265
setup = """## Setup
3366
3467
This rule requires that AWS Bedrock Integration be configured. For more information, see the AWS Bedrock integration documentation:

0 commit comments

Comments
 (0)