You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: AKS-Arc/telemetry-pod-resources.md
+18-15Lines changed: 18 additions & 15 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,6 @@
1
1
---
2
-
title: AKS Arc telemetry pod is consuming too much memory and CPU
3
-
description: Learn how to troubleshoot when AKS Arc telemetry pod is consuming too much memory and CPU
2
+
title: AKS Arc telemetry pod consumes too much memory and CPU
3
+
description: Learn how to troubleshoot when AKS Arc telemetry pod consumes too much memory and CPU.
4
4
ms.topic: troubleshooting
5
5
author: sethmanheim
6
6
ms.author: sethm
@@ -9,7 +9,7 @@ ms.reviewer: abha
9
9
10
10
---
11
11
12
-
# AKS Arc telemetry pod is consuming too much memory and CPU
12
+
# AKS Arc telemetry pod consumes too much memory and CPU
13
13
14
14
## Symptoms
15
15
@@ -26,20 +26,21 @@ NAME CPU(cores) MEMORY(bytes)
26
26
akshci-telemetry-5df56fd5-rjqk4 996m 152Mi
27
27
```
28
28
29
-
## Mitigation Steps
29
+
## Mitigation
30
30
31
-
To resolve this issue, we need to set default **resource limits** for the pods in the `kube-system` namespace.
31
+
To resolve this issue, set default **resource limits** for the pods in the `kube-system` namespace.
32
32
33
-
### Important Notes:
34
-
- Please verify if you have any pods in the **kube-system** namespace that might require more memory than the default limit setting below. If so, adjustments may be needed.
35
-
- The **LimitRange** is applied to the **namespace** (in this case, the `kube-system` namespace). The default resource limits will also apply to new pods that don't specify their own limits.
36
-
-**Existing pods** (including those that already have resource limits) will not be affected.
37
-
-**New pods** that don't specify their own resource limits will be constrained by the limits set below.
38
-
- After setting the resource limits and deleting the telemetry pod, the new pod might eventually hit the memory limit and get killed with **OOM (Out-Of-Memory)** errors. This is a **temporary mitigation**.
33
+
### Important notes
34
+
35
+
- Verify if you have any pods in the **kube-system** namespace that might require more memory than the default limit setting. If so, adjustments might be needed.
36
+
- The **LimitRange** is applied to the **namespace**; in this case, the `kube-system` namespace. The default resource limits also apply to new pods that don't specify their own limits.
37
+
-**Existing pods**, including those that already have resource limits, aren't affected.
38
+
-**New pods** that don't specify their own resource limits are constrained by the limits set in the next section.
39
+
- After you set the resource limits and delete the telemetry pod, the new pod might eventually hit the memory limit and generate **OOM (Out-Of-Memory)** errors. This is a temporary mitigation.
39
40
40
-
To proceed with setting the resource limits, you can run the following script. While the below script uses `az aksarc get-credentials`, you can also use `az connectedk8s proxy` to get the proxy kubeconfig and access the Kubernetes cluster.
41
+
To proceed with setting the resource limits, you can run the following script. While the script uses `az aksarc get-credentials`, you can also use `az connectedk8s proxy` to get the proxy kubeconfig and access the Kubernetes cluster.
41
42
42
-
### Define the **LimitRange YAML** to set default CPU and memory limits:
43
+
### Define the LimitRange YAML to set default CPU and memory limits
43
44
44
45
```powershell
45
46
# Set the $cluster_name and $resource_group of the aksarc cluster
@@ -75,14 +76,16 @@ sleep 5
75
76
kubectl get pods -l app=akshci-telemetry -n kube-system --kubeconfig "./kubeconfig-$cluster_name"
76
77
```
77
78
78
-
### Validate if the resource limits have been applied correctly
79
+
### Validate if the resource limits were applied correctly
79
80
80
81
1. Check the resource limits in the pod's YAML configuration:
0 commit comments