|
1 | 1 | ---
|
2 | 2 | title: Use auto-scaling in a Kubernetes cluster
|
3 |
| -description: Learn how to use Az CLI for cluster autoscaling. |
| 3 | +description: Learn how to use Azure CLI for cluster autoscaling. |
4 | 4 | ms.topic: how-to
|
5 | 5 | ms.custom: devx-track-azurecli
|
6 | 6 | author: sethmanheim
|
7 | 7 | ms.author: sethm
|
8 |
| -ms.date: 05/02/2025 |
| 8 | +ms.date: 06/09/2025 |
9 | 9 | ms.reviewer: abha
|
10 |
| -ms.lastreviewed: 05/02/2025 |
| 10 | +ms.lastreviewed: 06/09/2025 |
11 | 11 |
|
12 | 12 | # Intent: As a Kubernetes user, I want to use cluster autoscaling to grow my nodes to keep up with application demand.
|
13 | 13 | # Keyword: cluster autoscaling Kubernetes
|
@@ -147,16 +147,17 @@ az aksarc update \
|
147 | 147 | --cluster-autoscaler-profile ""
|
148 | 148 | ```
|
149 | 149 |
|
150 |
| -## Making effective use of autoscaler |
| 150 | +## Make effective use of autoscaler |
151 | 151 |
|
152 |
| -Now that the cluster and node pool are configured to automatically scale, you can optionally configure a workload to also scale in a way that makes use of the horizontal autoscaler capabilities. **Please note that the following guidance is not officially supported by Microsoft. It is shared as a best-effort recommendation based on open-source practices.** |
| 152 | +Now that the cluster and node pool are configured to automatically scale, you can optionally configure a workload to also scale in a way that makes use of the horizontal autoscaler capabilities. |
153 | 153 |
|
154 |
| -Two methods are available for workload scaling: |
155 |
| - |
156 |
| -* **Kubernetes Horizontal Pod Autoscaler**: Based on load characteristics, the Horizontal Pod Autoscaler (also known as the *horizontal autoscaler*) scales the pods of an application deployment to available nodes in the Kubernetes cluster. If no more nodes are available to be scheduled, the horizontal autoscaler instantiates a new node to which to schedule the pods. If application load goes down, the nodes are scaled back again. For the Horizontal Pod Autoscaler to work, **you must manually deploy the Metrics Server component in your AKS cluster**. For more information about horizontal pod autoscalar rules, see [Kubernetes horizontal pod autoscalar](https://kubernetes.io/docs/tasks/run-application/horizontal-pod-autoscale/). |
| 154 | +> [!NOTE] |
| 155 | +> The following guidance is not officially supported by Microsoft. It's shared as a best-effort recommendation based on open-source practices. |
157 | 156 |
|
158 |
| -* **Kubernetes node anti-affinity rules**: Anti-affinity rules for a Kubernetes deployment can specify that a set of pods can't be scaled on the same node, and a different node is required to scale the workload. In combination with either load characteristics or the number of target pods for the application instances, the horizontal autoscaler instantiates new nodes in the node pool to satisfy requests. If application demand subsides, the horizontal autoscaler scales down the node pool again. For more information about Kubernetes pod affinity rules, see [Assigning Pods to Nodes](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node). |
| 157 | +There are two methods available for workload scaling: |
159 | 158 |
|
| 159 | +- **Kubernetes Horizontal Pod Autoscaler**: Based on load characteristics, the Horizontal Pod Autoscaler (also known as the *horizontal autoscaler*) scales the pods of an application deployment to available nodes in the Kubernetes cluster. If no more nodes are available to be scheduled, the horizontal autoscaler instantiates a new node to which to schedule the pods. If the application load goes down, the nodes are scaled back again. For the Horizontal Pod Autoscaler to work, you must manually deploy the Metrics Server component in your AKS cluster. For more information about horizontal pod autoscaler rules, see [Kubernetes horizontal pod autoscaler](https://kubernetes.io/docs/tasks/run-application/horizontal-pod-autoscale/). |
| 160 | +- **Kubernetes node anti-affinity rules**: Anti-affinity rules for a Kubernetes deployment can specify that a set of pods can't be scaled on the same node, and a different node is required to scale the workload. In combination with either load characteristics or the number of target pods for the application instances, the horizontal autoscaler instantiates new nodes in the node pool to satisfy requests. If application demand subsides, the horizontal autoscaler scales down the node pool again. For more information about Kubernetes pod affinity rules, see [Assigning Pods to Nodes](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node). |
160 | 161 |
|
161 | 162 | ## Next steps
|
162 | 163 |
|
|
0 commit comments