You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/hdinsight/hdinsight-autoscale-clusters.md
+43-45Lines changed: 43 additions & 45 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -18,7 +18,14 @@ Azure HDInsight's free Autoscale feature can automatically increase or decrease
18
18
19
19
The Autoscale feature uses two types of conditions to trigger scaling events: thresholds for various cluster performance metrics (called *load-based scaling*) and time-based triggers (called *schedule-based scaling*). Load-based scaling changes the number of nodes in your cluster, within a range that you set, to ensure optimal CPU usage and minimize running cost. Schedule-based scaling changes the number of nodes in your cluster based on operations that you associate with specific dates and times.
20
20
21
-
### Metrics monitoring
21
+
### Choosing load-based or schedule-based scaling
22
+
23
+
Consider the following factors when choosing a scaling type:
24
+
25
+
* Load variance: does the load of the cluster follow a consistent pattern at specific times, on specific days? If not, load based scheduling is a better option.
26
+
* SLA requirements: Autoscale scaling is reactive instead of predictive. Will there be a sufficient delay between when the load starts to increase and when the cluster needs to be at its target size? If there are strict SLA requirements and the load is a fixed known pattern, 'schedule based' is a better option.
27
+
28
+
### Cluster metrics
22
29
23
30
Autoscale continuously monitors the cluster and collects the following metrics:
24
31
@@ -46,6 +53,24 @@ For scale-up, Autoscale issues a scale-up request to add the required number of
46
53
47
54
For scale-down, Autoscale issues a request to remove a certain number of nodes. The scale-down is based on the number of AM containers per node. And the current CPU and memory requirements. The service also detects which nodes are candidates for removal based on current job execution. The scale down operation first decommissions the nodes, and then removes them from the cluster.
48
55
56
+
### Cluster compatibility
57
+
58
+
> [!Important]
59
+
> The Azure HDInsight Autoscale feature was released for general availability on November 7th, 2019 for Spark and Hadoop clusters and included improvements not available in the preview version of the feature. If you created a Spark cluster prior to November 7th, 2019 and want to use the Autoscale feature on your cluster, the recommended path is to create a new cluster, and enable Autoscale on the new cluster.
60
+
>
61
+
> Autoscale for Interactive Query (LLAP) and HBase clusters is still in preview. Autoscale is only available on Spark, Hadoop, Interactive Query, and HBase clusters.
62
+
63
+
The following table describes the cluster types and versions that are compatible with the Autoscale feature.
64
+
65
+
| Version | Spark | Hive | LLAP | HBase | Kafka | Storm | ML |
66
+
|---|---|---|---|---|---|---|---|
67
+
| HDInsight 3.6 without ESP | Yes | Yes | Yes | Yes*| No | No | No |
68
+
| HDInsight 4.0 without ESP | Yes | Yes | Yes | Yes*| No | No | No |
69
+
| HDInsight 3.6 with ESP | Yes | Yes | Yes | Yes*| No | No | No |
70
+
| HDInsight 4.0 with ESP | Yes | Yes | Yes | Yes*| No | No | No |
71
+
72
+
\* HBase clusters can only be configured for schedule-based scaling, not load-based.
73
+
49
74
## Get started
50
75
51
76
### Create a cluster with load-based Autoscaling
@@ -94,24 +119,6 @@ Your subscription has a capacity quota for each region. The total number of core
94
119
95
120
For more information on HDInsight cluster creation using the Azure portal, see [Create Linux-based clusters in HDInsight using the Azure portal](hdinsight-hadoop-create-linux-clusters-portal.md).
96
121
97
-
## Cluster compatibility
98
-
99
-
> [!Important]
100
-
> The Azure HDInsight Autoscale feature was released for general availability on November 7th, 2019 for Spark and Hadoop clusters and included improvements not available in the preview version of the feature. If you created a Spark cluster prior to November 7th, 2019 and want to use the Autoscale feature on your cluster, the recommended path is to create a new cluster, and enable Autoscale on the new cluster.
101
-
>
102
-
> Autoscale for Interactive Query (LLAP) and HBase clusters is still in preview. Autoscale is only available on Spark, Hadoop, Interactive Query, and HBase clusters.
103
-
104
-
The following table describes the cluster types and versions that are compatible with the Autoscale feature.
105
-
106
-
| Version | Spark | Hive | LLAP | HBase | Kafka | Storm | ML |
107
-
|---|---|---|---|---|---|---|---|
108
-
| HDInsight 3.6 without ESP | Yes | Yes | Yes | Yes*| No | No | No |
109
-
| HDInsight 4.0 without ESP | Yes | Yes | Yes | Yes*| No | No | No |
110
-
| HDInsight 3.6 with ESP | Yes | Yes | Yes | Yes*| No | No | No |
111
-
| HDInsight 4.0 with ESP | Yes | Yes | Yes | Yes*| No | No | No |
112
-
113
-
\* HBase clusters can only be configured for schedule-based scaling, not load-based.
114
-
115
122
### Create a cluster with a Resource Manager template
116
123
117
124
#### Load-based autoscaling
@@ -198,32 +205,7 @@ Use the appropriate parameters in the request payload. The json payload below co
198
205
199
206
See the previous section on [enabling load-based autoscale](#load-based-autoscaling) for a full description of all payload parameters.
200
207
201
-
## Guidelines
202
-
203
-
### Choosing load-based or schedule-based scaling
204
-
205
-
Consider the following factors before making a decision on which mode to choose:
206
-
207
-
* Enable Autoscale during cluster creation.
208
-
* The minimum number of nodes should be at least three.
209
-
* Load variance: does the load of the cluster follow a consistent pattern at specific times, on specific days. If not, load based scheduling is a better option.
210
-
* SLA requirements: Autoscale scaling is reactive instead of predictive. Will there be a sufficient delay between when the load starts to increase and when the cluster needs to be at its target size? If there are strict SLA requirements and the load is a fixed known pattern, 'schedule based' is a better option.
211
-
212
-
### Consider the latency of scale up or scale down operations
213
-
214
-
It can take 10 to 20 minutes for a scaling operation to complete. When setting up a customized schedule, plan for this delay. For example, if you need the cluster size to be 20 at 9:00 AM, set the schedule trigger to an earlier time such as 8:30 AM so that the scaling operation has completed by 9:00 AM.
215
-
216
-
### Preparation for scaling down
217
-
218
-
During cluster scaling down process, Autoscale will decommission the nodes to meet the target size. If tasks are running on those nodes, Autoscale will wait until the tasks are completed. Since each worker node also serves a role in HDFS, the temp data will be shifted to the remaining nodes. So you should make sure there's enough space on the remaining nodes to host all the temp data.
219
-
220
-
The running jobs will continue. The pending jobs will wait for scheduling with fewer available worker nodes.
221
-
222
-
### Minimum cluster size
223
-
224
-
Don't scale your cluster down to fewer than three nodes. Scaling your cluster to fewer than three nodes can result in it getting stuck in safe mode because of insufficient file replication. For more information, see [Getting stuck in safe mode](./hdinsight-scaling-best-practices.md#getting-stuck-in-safe-mode).
225
-
226
-
## Monitoring
208
+
## Monitoring Autoscale activities
227
209
228
210
### Cluster status
229
211
@@ -251,6 +233,22 @@ Select **Metrics** under **Monitoring**. Then select **Add metric** and **Number
### Consider the latency of scale up or scale down operations
239
+
240
+
It can take 10 to 20 minutes for a scaling operation to complete. When setting up a customized schedule, plan for this delay. For example, if you need the cluster size to be 20 at 9:00 AM, set the schedule trigger to an earlier time such as 8:30 AM so that the scaling operation has completed by 9:00 AM.
241
+
242
+
### Preparation for scaling down
243
+
244
+
During cluster scaling down process, Autoscale will decommission the nodes to meet the target size. If tasks are running on those nodes, Autoscale will wait until the tasks are completed. Since each worker node also serves a role in HDFS, the temp data will be shifted to the remaining nodes. So you should make sure there's enough space on the remaining nodes to host all the temp data.
245
+
246
+
The running jobs will continue. The pending jobs will wait for scheduling with fewer available worker nodes.
247
+
248
+
### Minimum cluster size
249
+
250
+
Don't scale your cluster down to fewer than three nodes. Scaling your cluster to fewer than three nodes can result in it getting stuck in safe mode because of insufficient file replication. For more information, see [Getting stuck in safe mode](./hdinsight-scaling-best-practices.md#getting-stuck-in-safe-mode).
251
+
254
252
## Next steps
255
253
256
254
Read about guidelines for scaling clusters manually in [Scaling guidelines](hdinsight-scaling-best-practices.md)
0 commit comments