You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: content/en/ninja-workshops/12-alerting_monitoring_with_itsi/1-getting-started/1-access-cloud-instances.md
+10-2Lines changed: 10 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -4,6 +4,14 @@ linkTitle: 1.1 How to connect to your workshop environment
4
4
weight: 2
5
5
---
6
6
7
-
Access Show and sping up Workshop
7
+
# Starting up your Workshop
8
8
9
-
TBD
9
+
This workshop is available on Splunk Show and will take some time to start up all of your resources. It contains a Splunk environment with IT Service Intelligence, the Splunk Infrastructure Monitoring Add-On, as well as the recently updated AppDynamics Add-on all preconfigured.
10
+
11
+
The Workshop is titled **"Tech Summit 2025: OBS-122"** or you can go directly to it's entry on [Splunk Show](https://show.splunk.com/template/646/). It takes approximately 15 minutes to start up however data generation and ingestion will take up to a half hour.
Copy file name to clipboardExpand all lines: content/en/ninja-workshops/12-alerting_monitoring_with_itsi/2-creating-basic-alerts/_index.md
+25-35Lines changed: 25 additions & 35 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -12,66 +12,56 @@ This section covers the creation of basic alerts in Splunk Enterprise, AppDynami
12
12
13
13
Splunk alerts are triggered by search results that match specific criteria. We'll create a real-time alert that notifies us when a certain condition is met.
14
14
15
-
**Scenario:** Alert when the number of "error" events in the "application_logs" index exceeds 10 in the last 5 minutes.
15
+
**Scenario:** Alert when the number of "Invalid user" events in the "main" index exceeds 100 in the last 5 minutes.
16
16
17
17
**Steps:**
18
18
19
19
1.**Create a Search:** Start by creating a Splunk search that identifies the events you want to alert on. For example:
20
20
21
21
```splunk
22
-
index=application_logs level=error
22
+
index=main "Invalid user"
23
23
```
24
-
Use the time picker to select "Relative" and set the timespan to 10.
24
+
Use the time picker to select "Last 15 minutes"".
25
25
26
26
2.**Configure the Alert:**
27
27
* Click "Save As" and select "Alert."
28
-
* Give your alert a descriptive name (e.g., "Application Error Alert").
29
-
***Trigger Condition:**
30
-
***Scheduled:** Choose "Scheduled" to evaluate the search on a set schedule. Below scheduled will be the button to select the frequency, select "Run on Cron Schedule". If the time Range below that is different than 10 minutes, update it.
31
-
***Triggered when:** Select "Number of results" "is greater than" "10."
32
-
***Time Range:** Set to "5 minutes."
28
+
* Give your alert a descriptive name (e.g., "Numerous Invalid User Logins Attempted").
29
+
***Alert type:**
30
+
***Scheduled:** Choose "Scheduled" to evaluate the search on a set schedule. Below scheduled will be the button to select the frequency, select "Run on Cron Schedule".
31
+
***Cron Expression:***/15 * * * *
32
+
***Triggered when:** Select "Number of results" "is greater than" "100."
33
+
***Time Range:** Set to "15 minutes."
33
34
***Trigger Actions:**
34
35
* For this basic example, choose "Add to Triggered Alerts." In a real-world scenario, you'd configure email notifications, Slack integrations, or other actions.
35
36
***Save:** Save the alert.
36
37
37
-
**Explanation:** This alert runs the search every 10 minutes and triggers if the search returns more than 10 results. The "Add to Triggered Alerts" action simply adds a Alert to the Splunk Triggered Alerts list.
**Time Ranges and Frequency:** Since everything in Splunk core is a search, you need to consider the search timespan and frequency so that you are not a) searching the same data multiple times with an overlap timespan, b) missing events because of a gap between timespan and frequency, c) running too frequently and adding overhead or d) running too infrequently and experiencing delays in alerting.
40
-
41
-
42
-
## 2. AppDynamics Alerts (Health Rule Violations)
40
+
**Explanation:** This alert runs the search every 15 minutes and triggers if the search returns more than 100 results. The "Add to Triggered Alerts" action simply adds a Alert to the Splunk Triggered Alerts list.
43
41
44
-
**2. Create a Health Rule (or modify an existing one):**
45
-
* Click "Create Rule" (or edit an existing rule that applies to your application).
46
-
* Give the health rule a descriptive name (e.g., "Order Service Response Time Alert").
47
-
***Scope:** Select the application and tier (e.g., "OrderService").
48
-
***Conditions:**
49
-
* Choose the metric "Average Response Time."
50
-
* Set the threshold: "is greater than" "500" "milliseconds."
51
-
* Configure the evaluation frequency (how often AppDynamics checks the metric).
52
-
***Actions:**
53
-
* For this basic example, choose "Log to console." In a real-world scenario, you would configure email, SMS, or other notification channels.
54
-
***Save:** Save the health rule.
55
-
56
-
**Explanation:** This health rule continuously monitors the average response time of the "OrderService." If the response time exceeds 500ms, the health rule is violated, triggering the alert and the configured actions.
42
+
**Time Ranges and Frequency:** Since everything in Splunk core is a search, you need to consider the search timespan and frequency so that you are not a) searching the same data multiple times with an overlap timespan, b) missing events because of a gap between timespan and frequency, c) running too frequently and adding overhead or d) running too infrequently and experiencing delays in alerting.
* Give the detector a descriptive name (e.g., "High CPU Utilization Alert").
47
+
**Create a Detector:**
48
+
* Click "Detectors & SLOs" in the lefthand menu
49
+
* Click "Create Detector -> Custom Detector"
50
+
* Give the detector a descriptive name (e.g., "High CPU Utilization Alert - INITIALS").
64
51
***Signal:**
65
-
* Select the metric you want to monitor (e.g., "host.cpu.utilization"). Use the metric finder to locate the correct metric.
66
-
* Add any necessary filters to specify the host (e.g., `host:my-hostname`).
52
+
* Select the metric you want to monitor ("cpu.utilization").
53
+
* Add any necessary filters to specify the host (`service.name:otelshop-loadgenerator`).
54
+
* Click "Proceed to Alert Condition"
67
55
***Condition:**
68
-
*Set the threshold: "is above" "80" "%."
69
-
*Configure the evaluation frequency and the "for" duration (how long the condition must be true before the alert triggers).
56
+
*Select Static Threshold
57
+
*Set the threshold: "is above" "90"
70
58
***Notifications:**
71
59
* For this example, choose a simple notification method (e.g., a test webhook). In a real-world scenario, you would configure integrations with PagerDuty, Slack, or other notification systems.
72
60
***Save:** Save the detector.
73
61
74
-
**Explanation:** This detector monitors the CPU utilization metric for the specified host. If the CPU utilization exceeds 80% for the configured "for" duration, the detector triggers the alert and sends a notification.
**Explanation:** This detector monitors the CPU utilization metric for the specified service. If the CPU utilization exceeds 90% for the configured "for" duration, the detector triggers the alert and sends a notification.
# Starting with an Observability Cloud Based Service
8
+
9
+
1.**Access Services:** In ITSI click "Configuration", click on "Services".
10
+
11
+
2.**Create New Service: PaymentService2:** Click "Create New Service".
12
+
13
+
3.**Service Details (PaymentService2):**
14
+
***Title:** "PaymentService2"
15
+
***Description (Optional):** e.g., "Payment Service for Hipster Shop - version 2"
16
+
17
+
4.**Select Template:** Choose "Link service to a service template" and search for "Splunk APM Business Workflow KPIs" from the template dropdown. Click **Create** to save the new service.
18
+
19
+
6.**Entity Assignment:**
20
+
* The page will load and display the new Service and you will be on the Entities page. This demo defaults to selecting the *paymentservice:grpc.hipstershop.PaymentService/Charge* entity. In a real world situation you would need to match the workflow to the entity name manually.
21
+
***Direct Entity Selection (If Available):** Search for the entity using `sf_workflow="paymentservice:grpc.hipstershop.PaymentService/Charge"` and select it.
22
+
23
+
7.**Save Service (PaymentService2):** Click "Save" to create "PaymentService2".
24
+
25
+
8.**Settings:** Click the "Settings" tab, enable *Backfill* and keep that standard 7 days. Enable the Service, and click "Save"
26
+
27
+
## Setting PaymentService2's Service Health as a Dependency for Online-Boutique-US
28
+
29
+
1.**Locate Online-Boutique-US:** Find the "Online-Boutique-US" service in the service list.
30
+
31
+
2.**Edit Online-Boutique-US:** Click "Edit".
32
+
33
+
3.**Service Dependencies:** Look for the "Service Dependencies" section.
34
+
35
+
4.**Add Dependency:** There should be an option to add a dependent service. Search for "PaymentService2".
36
+
37
+
5.**Select KPI:** Check the box next to ServiceHealthScore for PaymentService2.
38
+
39
+
6.**Save Changes:** Save the changes to the "Online-Boutique-US" service.
40
+
41
+
## Verification
42
+
43
+
* Click on "Service Analyzer" and select the "Default Analyzer"
44
+
* Filter the service to just "Buttercup Business Health"
45
+
* Verify that *PaymentService2* is now present below *Online-Boutique-US* and should be in a grey status.
1.**Access Services:** In ITSI click "Configuration", click on "Services".
10
+
11
+
2.**Create Service: AD-Ecommerce2:** Click "Create Service -> Create Service".
12
+
13
+
3.**Service Details (AD-Ecommerce2):**
14
+
***Title:** "AD-Ecommerce2"
15
+
***Description (Optional):** e.g., "Ecommerce Service - version 2"
16
+
17
+
4.**Select Template:** Choose "Link service to a service template" and search for "AppDynamics App Performance Monitoring" from the template dropdown. Click **Create** to save the new service.
18
+
19
+
5.**Entity Assignment:**
20
+
* The page will load and display the new Service and you will be on the Entities page. This demo defaults to selecting the *AD-Ecommerce:18112:demo1.saas.appdynamics.com* entity. In a real world situation you would need to match the entity_name to the entity name manually.
21
+
***Direct Entity Selection (If Available):** Search for the entity using `entity_name="AD-Ecommerce:18112:demo1.saas.appdynamics.com"` and select it.
22
+
23
+
7.**Settings:** Click the "Settings" tab, enable *Backfill* and keep that standard 7 days. Enable the Service, and click "Save"
24
+
25
+
## Setting AD-Ecommerce2's Service Health as a Dependency for AD.Ecommerce
26
+
27
+
1.**Navigate back to Services page:** Click "Configuration -> Services"
28
+
29
+
2.**Locate AD.Ecommerce:** Find the "AD.Ecommerce" service in the service list.
30
+
31
+
3.**Edit AD.Ecommerce:** Click "Edit".
32
+
33
+
4.**Service Dependencies:** Look for the "Service Dependencies" section.
34
+
35
+
5.**Add Dependency:** There should be an option to add a dependent service. Search for "AD-Ecommerce2".
36
+
37
+
6.**Select KPI:** Check the box next to ServiceHealthScore for AD-Ecommerce2.
38
+
39
+
7.**Save Changes:** Save the changes to the "AD.Ecommerce" service.
40
+
41
+
## Verification
42
+
43
+
* Click on "Service Analyzer" and select the "Default Analyzer"
44
+
* Filter the service to just "Buttercup Business Health"
45
+
* Verify that *AD-Ecommerce2* is now present below *AD.Ecommerce* and should be in a grey status.
Copy file name to clipboardExpand all lines: content/en/ninja-workshops/12-alerting_monitoring_with_itsi/3-creating-services-in-itsi/_index.md
+7-2Lines changed: 7 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,5 +1,5 @@
1
1
---
2
-
title: Creating Basic Alerts
2
+
title: Creating Services in ITSI
3
3
linkTitle: 3. Creating Services in ITSI
4
4
weight: 1
5
5
---
@@ -10,5 +10,10 @@ This workshop outlines how to create a service in Splunk IT Service Intelligence
10
10
11
11
**Scenario:**
12
12
13
-
We have two existing services: "Astronomy Shop" (representing an application running in Kubernetes and being monitored by Splunk Observability Cloud) and "AD.ECommerce" (representing an application monitored by AppDynamics). We want to create a new service and add it as a dependent of one of those services. It is not necessary to create a service for both during your first run through this workshop so pick one that you are more interested in to start with.
13
+
We have two existing services: "Online-Boutique-US" (representing an application running in Kubernetes and being monitored by Splunk Observability Cloud) and "AD.ECommerce" (representing an application monitored by AppDynamics). We want to create a new service and add it as a dependent of one of those services. It is not necessary to create a service for both during your first run through this workshop so pick one that you are more interested in to start with.
This section guides you through configuring a basic alert in Splunk IT Service Intelligence (ITSI). We'll set up an alert that triggers when our previously created Service breaches a KPI threshold.
10
+
11
+
**Depending on the Service You Created, the KPI we use for this alert will change. In the instruction steps below replace Service Name and KPI appropriately**
12
+
13
+
***PaymentService2:** Business Workflow Error Rate
14
+
***AD-Ecommerce2:** Availability
15
+
16
+
**Steps:**
17
+
18
+
1.**Navigate to the KPI:**
19
+
* In ITSI, go to "Configuration" -> "Correlation Searches"
***Alert Fatigue:** Avoid setting up too many alerts or alerts with overly sensitive thresholds. This can lead to alert fatigue, where people become desensitized to alerts and might miss critical issues.
This section outlines the steps to create an aggregation policy in Splunk ITSI that matches the alerts we just set up. This policy will group related alerts, reducing noise and improving incident management.
10
+
11
+
**Depending on the Alert You Created, the title we use for this alert will change. In the instruction steps below replace AlertName with the Service Name used**
12
+
13
+
***PaymentService2** or
14
+
***AD-Ecommerce2**
15
+
16
+
## Steps
17
+
18
+
1.**Navigate to Notable Event Aggregation Policies:** In Splunk, go to "Configuration" -> "Notable Event Aggregation Policies".
19
+
20
+
2.**Create New Policy:** click the green "Create Notable Event Aggregation Policy" button in the upper right corner.
21
+
22
+
3.**Filtering Criteria:** This is the most important part. You'll define the criteria for alerts to be grouped by this policy. Click "Add Rule (OR)"
23
+
24
+
***Field:** Select "title" from the dropdown menu.
25
+
***Operator:** Choose "matches".
26
+
***Value:** Enter the string "*Service Name**". (make sure to include the *)
27
+
28
+
4.**Splitting Events:** Remove the "hosts" field that is provided by default and update it to use the "service" field. We want this generating new episodes for each Service that is found. In our example, it should only be 1.
29
+
30
+
5.**Breaking Criteria:** Configure how Episodes are broken or ended. We'll leave it as the default *"If an event occurs for which severity = Normal"*. Click Preview on the right to confirm it is picking up our Alert
31
+
32
+
6.**Click Next**
33
+
34
+
7.**Actions (Optional):** Define actions to be taken on aggregated alerts. For example, you can automatically create a ticket in ServiceNow or send an email notification. We're going to skip this part.
8.**Save Policy:** Click the "Next" button to create the aggregation policy.
43
+
44
+
## Verification
45
+
46
+
After saving the policy, navigate to the "Go to Episode Review" page and filter alerts for last 15 minutes and add a filter to status=New and search for our Service Name in the search box.
47
+
48
+
There may already be an episode named after our specific alert already, if so, close it out and wait for a new one to be generated with our new Title.
0 commit comments