Skip to content

Commit 5890a2c

Browse files
committed
Add Graphics and Steps
1 parent d10a5e7 commit 5890a2c

File tree

15 files changed

+208
-13
lines changed

15 files changed

+208
-13
lines changed

content/en/ninja-workshops/12-alerting_monitoring_with_itsi/1-getting-started/1-access-cloud-instances.md

Lines changed: 10 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -4,6 +4,14 @@ linkTitle: 1.1 How to connect to your workshop environment
44
weight: 2
55
---
66

7-
Access Show and sping up Workshop
7+
# Starting up your Workshop
88

9-
TBD
9+
This workshop is available on Splunk Show and will take some time to start up all of your resources. It contains a Splunk environment with IT Service Intelligence, the Splunk Infrastructure Monitoring Add-On, as well as the recently updated AppDynamics Add-on all preconfigured.
10+
11+
The Workshop is titled **"Tech Summit 2025: OBS-122"** or you can go directly to it's entry on [Splunk Show](https://show.splunk.com/template/646/). It takes approximately 15 minutes to start up however data generation and ingestion will take up to a half hour.
12+
13+
![show-entry](../images/show_entry.png?classes=inline)
14+
15+
# Splunk Observability Cloud Access
16+
17+
Creating an alert in Observability Cloud should be done in the Observability Cloud **Show Playground** Org.

content/en/ninja-workshops/12-alerting_monitoring_with_itsi/2-creating-basic-alerts/_index.md

Lines changed: 15 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -12,29 +12,31 @@ This section covers the creation of basic alerts in Splunk Enterprise, AppDynami
1212

1313
Splunk alerts are triggered by search results that match specific criteria. We'll create a real-time alert that notifies us when a certain condition is met.
1414

15-
**Scenario:** Alert when the number of "error" events in the "application_logs" index exceeds 10 in the last 5 minutes.
15+
**Scenario:** Alert when the number of "Invalid user" events in the "main" index exceeds 100 in the last 5 minutes.
1616

1717
**Steps:**
1818

1919
1. **Create a Search:** Start by creating a Splunk search that identifies the events you want to alert on. For example:
2020

2121
```splunk
22-
index=application_logs level=error
22+
index=main "Invalid user"
2323
```
2424
Use the time picker to select "Relative" and set the timespan to 10.
2525

2626
2. **Configure the Alert:**
2727
* Click "Save As" and select "Alert."
28-
* Give your alert a descriptive name (e.g., "Application Error Alert").
28+
* Give your alert a descriptive name (e.g., "Numerous Invalid User Logins Attempted").
2929
* **Trigger Condition:**
3030
* **Scheduled:** Choose "Scheduled" to evaluate the search on a set schedule. Below scheduled will be the button to select the frequency, select "Run on Cron Schedule". If the time Range below that is different than 10 minutes, update it.
31-
* **Triggered when:** Select "Number of results" "is greater than" "10."
31+
* **Triggered when:** Select "Number of results" "is greater than" "100."
3232
* **Time Range:** Set to "5 minutes."
3333
* **Trigger Actions:**
3434
* For this basic example, choose "Add to Triggered Alerts." In a real-world scenario, you'd configure email notifications, Slack integrations, or other actions.
3535
* **Save:** Save the alert.
3636

37-
**Explanation:** This alert runs the search every 10 minutes and triggers if the search returns more than 10 results. The "Add to Triggered Alerts" action simply adds a Alert to the Splunk Triggered Alerts list.
37+
![show-entry](../images/save_as_alert.png?classes=inline)
38+
39+
**Explanation:** This alert runs the search every 10 minutes and triggers if the search returns more than 100 results. The "Add to Triggered Alerts" action simply adds a Alert to the Splunk Triggered Alerts list.
3840

3941
**Time Ranges and Frequency:** Since everything in Splunk core is a search, you need to consider the search timespan and frequency so that you are not a) searching the same data multiple times with an overlap timespan, b) missing events because of a gap between timespan and frequency, c) running too frequently and adding overhead or d) running too infrequently and experiencing delays in alerting.
4042

@@ -56,22 +58,25 @@ Splunk alerts are triggered by search results that match specific criteria. We'l
5658
**Explanation:** This health rule continuously monitors the average response time of the "OrderService." If the response time exceeds 500ms, the health rule is violated, triggering the alert and the configured actions.
5759

5860

59-
## 3. Splunk Observability Cloud Alerts (Detectors) (Continued)
61+
## 3. Splunk Observability Cloud Alerts (Detectors)
6062

6163
**2. Create a Detector:**
64+
* Click "Detectors & SLOs" in the lefthand menu
6265
* Click "Create Detector."
6366
* Give the detector a descriptive name (e.g., "High CPU Utilization Alert").
6467
* **Signal:**
65-
* Select the metric you want to monitor (e.g., "host.cpu.utilization"). Use the metric finder to locate the correct metric.
66-
* Add any necessary filters to specify the host (e.g., `host:my-hostname`).
68+
* Select the metric you want to monitor (e.g., "cpu.utilization"). Use the metric finder to locate the correct metric.
69+
* Add any necessary filters to specify the host (e.g., `service.name:otelshop-loadgenerator`).
6770
* **Condition:**
68-
* Set the threshold: "is above" "80" "%."
71+
* Set the threshold: "is above" "90" "%."
6972
* Configure the evaluation frequency and the "for" duration (how long the condition must be true before the alert triggers).
7073
* **Notifications:**
7174
* For this example, choose a simple notification method (e.g., a test webhook). In a real-world scenario, you would configure integrations with PagerDuty, Slack, or other notification systems.
7275
* **Save:** Save the detector.
7376

74-
**Explanation:** This detector monitors the CPU utilization metric for the specified host. If the CPU utilization exceeds 80% for the configured "for" duration, the detector triggers the alert and sends a notification.
77+
![show-entry](../images/detector_preview.png?classes=inline)
78+
79+
**Explanation:** This detector monitors the CPU utilization metric for the specified service. If the CPU utilization exceeds 90% for the configured "for" duration, the detector triggers the alert and sends a notification.
7580

7681
**Important Considerations for All Platforms:**
7782

content/en/ninja-workshops/12-alerting_monitoring_with_itsi/3-creating-services-in-itsi/1-creating-o11y-service.md

Whitespace-only changes.
Lines changed: 47 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,47 @@
1+
---
2+
title: Creating an AppD Based Service
3+
linkTitle: 3.2 Creating an AppD Based Service
4+
weight: 3
5+
---
6+
7+
# Starting with an AppDynamics Based Service
8+
9+
1. **Access Services:** In ITSI click "Configuration", click on "Services".
10+
11+
2. **Create New Service: AD-Ecommerce2:** Click "Create New Service".
12+
13+
3. **Service Details (AD-Ecommerce2):**
14+
* **Title:** "AD-Ecommerce2"
15+
* **Description (Optional):** e.g., "Ecommerce Service - version 2"
16+
17+
4. **Select Template:** Choose "Link service to a service template" and search for "AppDynamics App Performance Monitoring" from the template dropdown. Click **Create** to save the new service.
18+
19+
6. **Entity Assignment:**
20+
* The page will load and display the new Service and you will be on the Entities page. This demo defaults to selecting the *AD-Ecommerce:18112:demo1.saas.appdynamics.com* entity. In a real world situation you would need to match the entity_name to the entity name manually.
21+
* **Direct Entity Selection (If Available):** Search for the entity using `entity_name="AD-Ecommerce:18112:demo1.saas.appdynamics.com"` and select it.
22+
23+
7. **Save Service (AD-Ecommerce2):** Click "Save" to create "AD-Ecommerce2".
24+
25+
8. **Settings:** Click the "Settings" tab, enable *Backfill* and keep that standard 7 days. Enable the Service, and click "Save"
26+
27+
## Setting AD-Ecommerce2's Service Health as a Dependency for AD.Ecommerce
28+
29+
1. **Locate AD.Ecommerce:** Find the "AD.Ecommerce" service in the service list.
30+
31+
2. **Edit AD.Ecommerce:** Click "Edit".
32+
33+
3. **Service Dependencies:** Look for the "Service Dependencies" section.
34+
35+
4. **Add Dependency:** There should be an option to add a dependent service. Search for "AD-Ecommerce2".
36+
37+
5. **Select KPI:** Check the box next to ServiceHealthScore for AD-Ecommerce2.
38+
39+
6. **Save Changes:** Save the changes to the "AD.Ecommerce" service.
40+
41+
## Verification
42+
43+
* Click on "Service Analyzer" and select the "Default Analyzer"
44+
* Filter the service to just "Buttercup Business Health"
45+
* Verify that *AD-Ecommerce2* is now present below *AD.Ecommerce* and should be in a grey status.
46+
47+
![show-entry](../images/service_tree_appd.png?classes=inline)

content/en/ninja-workshops/12-alerting_monitoring_with_itsi/3-creating-services-in-itsi/_index.md

Lines changed: 39 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,5 +10,43 @@ This workshop outlines how to create a service in Splunk IT Service Intelligence
1010

1111
**Scenario:**
1212

13-
We have two existing services: "Astronomy Shop" (representing an application running in Kubernetes and being monitored by Splunk Observability Cloud) and "AD.ECommerce" (representing an application monitored by AppDynamics). We want to create a new service and add it as a dependent of one of those services. It is not necessary to create a service for both during your first run through this workshop so pick one that you are more interested in to start with.
13+
We have two existing services: "Online-Boutique-US" (representing an application running in Kubernetes and being monitored by Splunk Observability Cloud) and "AD.ECommerce" (representing an application monitored by AppDynamics). We want to create a new service and add it as a dependent of one of those services. It is not necessary to create a service for both during your first run through this workshop so pick one that you are more interested in to start with.
1414

15+
![show-entry](../images/service_tree_start.png?classes=inline)
16+
17+
## Starting with an Observability Cloud Based Service
18+
19+
1. **Access Services:** In ITSI click "Configuration", click on "Services".
20+
21+
2. **Create New Service: PaymentService2:** Click "Create New Service".
22+
23+
3. **Service Details (PaymentService2):**
24+
* **Title:** "PaymentService2"
25+
* **Description (Optional):** e.g., "Payment Service for Hipster Shop - version 2"
26+
27+
4. **Select Template:** Choose "Link service to a service template" and search for "Splunk APM Business Workflow KPIs" from the template dropdown. Click **Create** to save the new service.
28+
29+
6. **Entity Assignment:**
30+
* The page will load and display the new Service and you will be on the Entities page. This demo defaults to selecting the *paymentservice:grpc.hipstershop.PaymentService/Charge* entity. In a real world situation you would need to match the workflow to the entity name manually.
31+
* **Direct Entity Selection (If Available):** Search for the entity using `sf_workflow="paymentservice:grpc.hipstershop.PaymentService/Charge"` and select it.
32+
33+
7. **Save Service (PaymentService2):** Click "Save" to create "PaymentService2".
34+
35+
## Setting PaymentService2's Service Health as a Dependency for Online-Boutique-US
36+
37+
1. **Locate Online-Boutique-US:** Find the "Online-Boutique-US" service in the service list.
38+
39+
2. **Edit Online-Boutique-US:** Click "Edit".
40+
41+
3. **Service Dependencies:** Look for the "Service Dependencies" section.
42+
43+
4. **Add Dependency:** There should be an option to add a dependent service. Search for "PaymentService2".
44+
45+
5. **Select KPI:** Check the box next to ServiceHealthScore for PaymentService2.
46+
47+
6. **Save Changes:** Save the changes to the "Online-Boutique-US" service.
48+
49+
## Verification
50+
51+
* Verify "PaymentService2" is created and linked to the correct entity.
52+
* Verify "Online-Boutique-US" now has "PaymentService2" (specifically its Service Health Score) as a dependency. Changes in the health of "PaymentService2" should now impact the health of "Online-Boutique-US."
Lines changed: 46 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,46 @@
1+
---
2+
title: Creating Alerts in ITSI
3+
linkTitle: 4. Creating Alerts in ITSI
4+
weight: 1
5+
---
6+
7+
# Configuring a Basic Alert in Splunk ITSI
8+
9+
This section guides you through configuring a basic alert in Splunk IT Service Intelligence (ITSI). We'll set up an alert that triggers when our previously created Service breaches a KPI threshold.
10+
11+
**Depending on the Service You Created, the KPI we use for this alert will change. In the instruction steps below replace Service Name and KPI appropriately**
12+
13+
* **PaymentService2:** Business Workflow Error Rate
14+
* **AD-Ecommerce2:** Availability
15+
16+
**Steps:**
17+
18+
1. **Navigate to the KPI:**
19+
* In ITSI, go to "Configuration" -> "Correlation Searches"
20+
* Click "Create New Search"
21+
22+
2. **Configure the new search:**
23+
* **Search Title:** *Service Name* *KPI* Critical
24+
* **Description:** *Service Name* *KPI* Critical
25+
* **Search:**
26+
```
27+
index=itsi_summary kpi="*KPI*" alert_severity=critical
28+
```
29+
* **Time Range:** Last 5 minutes
30+
* **Service:** *Service Name*
31+
* **Entity Lookup Field:** itsi_service_id
32+
* **Run Every:** 5 minutes
33+
* **Notable Event Title:** *Service Name* *KPI* Critical
34+
* **Notable Event Identified Fields:** source
35+
36+
![show-entry](../images/alerts.png?classes=inline)
37+
38+
**After Creating the Alert:**
39+
40+
* You will need to wait 5-10 minutes for the alert to run
41+
* The alert will be listed in the "Alerts and Episodes" Pane in ITSI.
42+
43+
44+
**Important Considerations:**
45+
46+
* **Alert Fatigue:** Avoid setting up too many alerts or alerts with overly sensitive thresholds. This can lead to alert fatigue, where people become desensitized to alerts and might miss critical issues.
Lines changed: 51 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,51 @@
1+
---
2+
title: Creating Episodes in ITSI
3+
linkTitle: 5. Creating Episodes in ITSI
4+
weight: 1
5+
---
6+
7+
# Creating an Aggregation Policy in Splunk ITSI
8+
9+
This section outlines the steps to create an aggregation policy in Splunk ITSI that matches the alerts we just set up. This policy will group related alerts, reducing noise and improving incident management.
10+
11+
**Depending on the Alert You Created, the title we use for this alert will change. In the instruction steps below replace AlertName with the Service Name used**
12+
13+
* **PaymentService2** or
14+
* **AD-Ecommerce2**
15+
16+
## Steps
17+
18+
1. **Navigate to Notable Event Aggregation Policies:** In Splunk, go to the Notable Event Aggregation Policies section (usually under "Configuration" -> "Notable Event Aggregation Policies").
19+
20+
2. **Create New Policy:** click the green "Create Notable Event Aggregation Policy" button in the upper right corner.
21+
22+
3. **Filtering Criteria:** This is the most important part. You'll define the criteria for alerts to be grouped by this policy. Click "Add Rule (OR)"
23+
24+
* **Field:** Select "title" from the dropdown menu.
25+
* **Operator:** Choose "matches".
26+
* **Value:** Enter the string "*Service Name**".
27+
28+
4. **Splitting Events:** Remove the "hosts" field that is provided by default and update it to use the "service" field. We want this generating new episodes for each Service that is found. In our example, it should only be 1.
29+
30+
5. **Breaking Criteria:** Configure how Episodes are broken or ended. We'll leave it as the default *"If an event occurs for which severity = Normal"*. Click Preview on the right to confirm it is picking up our Alert
31+
32+
6. **Click Next**
33+
34+
7. **Actions (Optional):** Define actions to be taken on aggregated alerts. For example, you can automatically create a ticket in ServiceNow or send an email notification. We're going to skip this part.
35+
36+
8. **Click Next**
37+
38+
9. **Policy Name and Description:**
39+
* **Policy Title:** *Service Name* Alert Grouping
40+
* **Description:** Grouping *Service Name* alerts together.
41+
42+
8. **Save Policy:** Click the "Save" button to create the aggregation policy.
43+
44+
## Verification
45+
46+
After saving the policy, navigate to the "Alerts and Episodes" page and filter alerts for last 15 minutes and add a filter to status=New and search for our Service Name in the search box.
47+
48+
There may already be an episode named after our specific alert already, if so, close it out and wait for a new one to be generated with our new Title.
49+
50+
![show-entry](../images/episode.png?classes=inline)
51+
90.1 KB
Loading
121 KB
Loading
77.9 KB
Loading

0 commit comments

Comments
 (0)