Skip to content

Commit d831171

Browse files
committed
reorganizing autoscale doc
1 parent 03e1e52 commit d831171

File tree

1 file changed

+43
-45
lines changed

1 file changed

+43
-45
lines changed

articles/hdinsight/hdinsight-autoscale-clusters.md

Lines changed: 43 additions & 45 deletions
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,14 @@ Azure HDInsight's free Autoscale feature can automatically increase or decrease
1818

1919
The Autoscale feature uses two types of conditions to trigger scaling events: thresholds for various cluster performance metrics (called *load-based scaling*) and time-based triggers (called *schedule-based scaling*). Load-based scaling changes the number of nodes in your cluster, within a range that you set, to ensure optimal CPU usage and minimize running cost. Schedule-based scaling changes the number of nodes in your cluster based on operations that you associate with specific dates and times.
2020

21-
### Metrics monitoring
21+
### Choosing load-based or schedule-based scaling
22+
23+
Consider the following factors when choosing a scaling type:
24+
25+
* Load variance: does the load of the cluster follow a consistent pattern at specific times, on specific days? If not, load based scheduling is a better option.
26+
* SLA requirements: Autoscale scaling is reactive instead of predictive. Will there be a sufficient delay between when the load starts to increase and when the cluster needs to be at its target size? If there are strict SLA requirements and the load is a fixed known pattern, 'schedule based' is a better option.
27+
28+
### Cluster metrics
2229

2330
Autoscale continuously monitors the cluster and collects the following metrics:
2431

@@ -46,6 +53,24 @@ For scale-up, Autoscale issues a scale-up request to add the required number of
4653

4754
For scale-down, Autoscale issues a request to remove a certain number of nodes. The scale-down is based on the number of AM containers per node. And the current CPU and memory requirements. The service also detects which nodes are candidates for removal based on current job execution. The scale down operation first decommissions the nodes, and then removes them from the cluster.
4855

56+
### Cluster compatibility
57+
58+
> [!Important]
59+
> The Azure HDInsight Autoscale feature was released for general availability on November 7th, 2019 for Spark and Hadoop clusters and included improvements not available in the preview version of the feature. If you created a Spark cluster prior to November 7th, 2019 and want to use the Autoscale feature on your cluster, the recommended path is to create a new cluster, and enable Autoscale on the new cluster.
60+
>
61+
> Autoscale for Interactive Query (LLAP) and HBase clusters is still in preview. Autoscale is only available on Spark, Hadoop, Interactive Query, and HBase clusters.
62+
63+
The following table describes the cluster types and versions that are compatible with the Autoscale feature.
64+
65+
| Version | Spark | Hive | LLAP | HBase | Kafka | Storm | ML |
66+
|---|---|---|---|---|---|---|---|
67+
| HDInsight 3.6 without ESP | Yes | Yes | Yes | Yes* | No | No | No |
68+
| HDInsight 4.0 without ESP | Yes | Yes | Yes | Yes* | No | No | No |
69+
| HDInsight 3.6 with ESP | Yes | Yes | Yes | Yes* | No | No | No |
70+
| HDInsight 4.0 with ESP | Yes | Yes | Yes | Yes* | No | No | No |
71+
72+
\* HBase clusters can only be configured for schedule-based scaling, not load-based.
73+
4974
## Get started
5075

5176
### Create a cluster with load-based Autoscaling
@@ -94,24 +119,6 @@ Your subscription has a capacity quota for each region. The total number of core
94119
95120
For more information on HDInsight cluster creation using the Azure portal, see [Create Linux-based clusters in HDInsight using the Azure portal](hdinsight-hadoop-create-linux-clusters-portal.md).
96121

97-
## Cluster compatibility
98-
99-
> [!Important]
100-
> The Azure HDInsight Autoscale feature was released for general availability on November 7th, 2019 for Spark and Hadoop clusters and included improvements not available in the preview version of the feature. If you created a Spark cluster prior to November 7th, 2019 and want to use the Autoscale feature on your cluster, the recommended path is to create a new cluster, and enable Autoscale on the new cluster.
101-
>
102-
> Autoscale for Interactive Query (LLAP) and HBase clusters is still in preview. Autoscale is only available on Spark, Hadoop, Interactive Query, and HBase clusters.
103-
104-
The following table describes the cluster types and versions that are compatible with the Autoscale feature.
105-
106-
| Version | Spark | Hive | LLAP | HBase | Kafka | Storm | ML |
107-
|---|---|---|---|---|---|---|---|
108-
| HDInsight 3.6 without ESP | Yes | Yes | Yes | Yes* | No | No | No |
109-
| HDInsight 4.0 without ESP | Yes | Yes | Yes | Yes* | No | No | No |
110-
| HDInsight 3.6 with ESP | Yes | Yes | Yes | Yes* | No | No | No |
111-
| HDInsight 4.0 with ESP | Yes | Yes | Yes | Yes* | No | No | No |
112-
113-
\* HBase clusters can only be configured for schedule-based scaling, not load-based.
114-
115122
### Create a cluster with a Resource Manager template
116123

117124
#### Load-based autoscaling
@@ -198,32 +205,7 @@ Use the appropriate parameters in the request payload. The json payload below co
198205

199206
See the previous section on [enabling load-based autoscale](#load-based-autoscaling) for a full description of all payload parameters.
200207

201-
## Guidelines
202-
203-
### Choosing load-based or schedule-based scaling
204-
205-
Consider the following factors before making a decision on which mode to choose:
206-
207-
* Enable Autoscale during cluster creation.
208-
* The minimum number of nodes should be at least three.
209-
* Load variance: does the load of the cluster follow a consistent pattern at specific times, on specific days. If not, load based scheduling is a better option.
210-
* SLA requirements: Autoscale scaling is reactive instead of predictive. Will there be a sufficient delay between when the load starts to increase and when the cluster needs to be at its target size? If there are strict SLA requirements and the load is a fixed known pattern, 'schedule based' is a better option.
211-
212-
### Consider the latency of scale up or scale down operations
213-
214-
It can take 10 to 20 minutes for a scaling operation to complete. When setting up a customized schedule, plan for this delay. For example, if you need the cluster size to be 20 at 9:00 AM, set the schedule trigger to an earlier time such as 8:30 AM so that the scaling operation has completed by 9:00 AM.
215-
216-
### Preparation for scaling down
217-
218-
During cluster scaling down process, Autoscale will decommission the nodes to meet the target size. If tasks are running on those nodes, Autoscale will wait until the tasks are completed. Since each worker node also serves a role in HDFS, the temp data will be shifted to the remaining nodes. So you should make sure there's enough space on the remaining nodes to host all the temp data.
219-
220-
The running jobs will continue. The pending jobs will wait for scheduling with fewer available worker nodes.
221-
222-
### Minimum cluster size
223-
224-
Don't scale your cluster down to fewer than three nodes. Scaling your cluster to fewer than three nodes can result in it getting stuck in safe mode because of insufficient file replication. For more information, see [Getting stuck in safe mode](./hdinsight-scaling-best-practices.md#getting-stuck-in-safe-mode).
225-
226-
## Monitoring
208+
## Monitoring Autoscale activities
227209

228210
### Cluster status
229211

@@ -251,6 +233,22 @@ Select **Metrics** under **Monitoring**. Then select **Add metric** and **Number
251233

252234
![Enable worker node schedule-based autoscale metric](./media/hdinsight-autoscale-clusters/hdinsight-autoscale-clusters-chart-metric.png)
253235

236+
## Other considerations
237+
238+
### Consider the latency of scale up or scale down operations
239+
240+
It can take 10 to 20 minutes for a scaling operation to complete. When setting up a customized schedule, plan for this delay. For example, if you need the cluster size to be 20 at 9:00 AM, set the schedule trigger to an earlier time such as 8:30 AM so that the scaling operation has completed by 9:00 AM.
241+
242+
### Preparation for scaling down
243+
244+
During cluster scaling down process, Autoscale will decommission the nodes to meet the target size. If tasks are running on those nodes, Autoscale will wait until the tasks are completed. Since each worker node also serves a role in HDFS, the temp data will be shifted to the remaining nodes. So you should make sure there's enough space on the remaining nodes to host all the temp data.
245+
246+
The running jobs will continue. The pending jobs will wait for scheduling with fewer available worker nodes.
247+
248+
### Minimum cluster size
249+
250+
Don't scale your cluster down to fewer than three nodes. Scaling your cluster to fewer than three nodes can result in it getting stuck in safe mode because of insufficient file replication. For more information, see [Getting stuck in safe mode](./hdinsight-scaling-best-practices.md#getting-stuck-in-safe-mode).
251+
254252
## Next steps
255253

256254
Read about guidelines for scaling clusters manually in [Scaling guidelines](hdinsight-scaling-best-practices.md)

0 commit comments

Comments
 (0)