Skip to content

Commit ccfdb48

Browse files
Merge pull request #278532 from vdiec/patch-1
Update kubernetes-metric-alerts.md
2 parents b56346e + 96e391d commit ccfdb48

File tree

1 file changed

+14
-10
lines changed

1 file changed

+14
-10
lines changed

articles/azure-monitor/containers/kubernetes-metric-alerts.md

Lines changed: 14 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -2,8 +2,8 @@
22
title: Recommended alert rules for Kubernetes clusters
33
description: Describes how to enable recommended metric alerts rules for a Kubernetes cluster in Azure Monitor.
44
ms.topic: conceptual
5-
ms.date: 03/05/2024
6-
ms.reviewer: aul
5+
ms.date: 06/17/2024
6+
ms.reviewer: vdiec
77
---
88

99
# Recommended alert rules for Kubernetes clusters
@@ -20,6 +20,10 @@ There are two types of metric alert rules used with Kubernetes clusters.
2020
## Enable recommended alert rules
2121
Use one of the following methods to enable the recommended alert rules for your cluster. You can enable both Prometheus and platform metric alert rules for the same cluster.
2222

23+
>[!NOTE]
24+
>To enable recommended alerts on Arc-enabled Kubernetes clusters, ARM templates are the only supported method.
25+
>
26+
2327
### [Azure portal](#tab/portal)
2428
Using the Azure portal, the Prometheus rule group will be created in the same region as the cluster.
2529

@@ -136,7 +140,7 @@ The following tables list the details of each recommended alert rule. Source cod
136140
|:---|:---|:---:|:---:|
137141
| KubeCPUQuotaOvercommit | The CPU resource quota allocated to namespaces exceeds the available CPU resources on the cluster's nodes by more than 50% for the last 5 minutes. | >1.5 | 5 |
138142
| KubeMemoryQuotaOvercommit | The memory resource quota allocated to namespaces exceeds the available memory resources on the cluster's nodes by more than 50% for the last 5 minutes. | >1.5 | 5 |
139-
| Number of OOM killed containers is greater than 0 | One or more containers within pods have been killed due to out-of-memory (OOM) events for the last 5 minutes. | >0 | 5 |
143+
| KubeContainerOOMKilledCount | One or more containers within pods have been killed due to out-of-memory (OOM) events for the last 5 minutes. | >0 | 5 |
140144
| KubeClientErrors | The rate of client errors (HTTP status codes starting with 5xx) in Kubernetes API requests exceeds 1% of the total API request rate for the last 15 minutes. | >0.01 | 15 |
141145
| KubePersistentVolumeFillingUp | The persistent volume is filling up and is expected to run out of available space evaluated on the available space ratio, used space, and predicted linear trend of available space over the last 6 hours. These conditions are evaluated over the last 60 minutes. | N/A | 60 |
142146
| KubePersistentVolumeInodesFillingUp | Less than 3% of the inodes within a persistent volume are available for the last 15 minutes. | <0.03 | 15 |
@@ -158,29 +162,29 @@ The following tables list the details of each recommended alert rule. Source cod
158162

159163
| Alert name | Description | Default threshold | Timeframe (minutes) |
160164
|:---|:---|:---:|:---:|
161-
| Average PV usage is greater than 80% | The average usage of Persistent Volumes (PVs) on pod exceeds 80% for the last 15 minutes. | >0.8 | 15 |
165+
| KubePVUsageHigh | The average usage of Persistent Volumes (PVs) on pod exceeds 80% for the last 15 minutes. | >0.8 | 15 |
162166
| KubeDeploymentReplicasMismatch | There is a mismatch between the desired number of replicas and the number of available replicas for the last 10 minutes. | N/A | 10 |
163167
| KubeStatefulSetReplicasMismatch | The number of ready replicas in the StatefulSet does not match the total number of replicas in the StatefulSet for the last 15 minutes. | N/A | 15 |
164168
| KubeHpaReplicasMismatch | The Horizontal Pod Autoscaler in the cluster has not matched the desired number of replicas for the last 15 minutes. | N/A | 15 |
165169
| KubeHpaMaxedOut | The Horizontal Pod Autoscaler (HPA) in the cluster has been running at the maximum replicas for the last 15 minutes. | N/A | 15 |
166170
| KubePodCrashLooping | One or more pods is in a CrashLoopBackOff condition, where the pod continuously crashes after startup and fails to recover successfully for the last 15 minutes. | >=1 | 15 |
167171
| KubeJobStale | At least one Job instance did not complete successfully for the last 6 hours. | >0 | 360 |
168-
| Pod container restarted in last 1 hour | One or more containers within pods in the Kubernetes cluster have been restarted at least once within the last hour. | >0 | 15 |
169-
| Ready state of pods is less than 80% | The percentage of pods in a ready state falls below 80% for any deployment or daemonset in the Kubernetes cluster for the last 5 minutes. | <0.8 | 5 |
170-
| Number of pods in failed state are greater than 0. | One or more pods is in a failed state for the last 5 minutes. | >0 | 5 |
172+
| KubePodContainerRestart | One or more containers within pods in the Kubernetes cluster have been restarted at least once within the last hour. | >0 | 15 |
173+
| KubePodReadyStateLow | The percentage of pods in a ready state falls below 80% for any deployment or daemonset in the Kubernetes cluster for the last 5 minutes. | <0.8 | 5 |
174+
| KubePodFailedState | One or more pods is in a failed state for the last 5 minutes. | >0 | 5 |
171175
| KubePodNotReadyByController | One or more pods are not in a ready state (i.e., in the "Pending" or "Unknown" phase) for the last 15 minutes. | >0 | 15 |
172176
| KubeStatefulSetGenerationMismatch | The observed generation of a Kubernetes StatefulSet does not match its metadata generation for the last 15 minutes. | N/A | 15 |
173177
| KubeJobFailed | One or more Kubernetes jobs have failed within the last 15 minutes. | >0 | 15 |
174-
| Average CPU usage per container is greater than 95% | The average CPU usage per container exceeds 95% for the last 5 minutes. | >0.95 | 5 |
175-
| Average Memory usage per container is greater than 95% | The average memory usage per container exceeds 95% for the last 5 minutes. | >0.95 | 10 |
178+
| KubeContainerAverageCPUHigh | The average CPU usage per container exceeds 95% for the last 5 minutes. | >0.95 | 5 |
179+
| KubeContainerAverageMemoryHigh | The average memory usage per container exceeds 95% for the last 5 minutes. | >0.95 | 10 |
176180
| KubeletPodStartUpLatencyHigh | The 99th percentile of the pod startup latency exceeds 60 seconds for the last 10 minutes. | >60 | 10 |
177181

178182
### Platform metric alert rules
179183

180184
| Alert name | Description | Default threshold | Timeframe (minutes) |
181185
|:---|:---|:---:|:---:|
182186
| Node cpu percentage is greater than 95% | The node CPU percentage is greater than 95% for the last 5 minutes. | 95 | 5 |
183-
| Node memory working set percentage is greater than 100% | The node memory working set percentage is greater than 95% for the last 5 minutes. | 100 | 5 |
187+
| Node memory working set percentage is greater than 100% | The node memory working set percentage is greater than 100% for the last 5 minutes. | 100 | 5 |
184188

185189

186190
## Legacy Container insights metric alerts (preview)

0 commit comments

Comments
 (0)