You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This article walks you through setup in the [Azure portal](https://portal.azure.com), where you can create an HDInsight cluster.
36
+
This article walks you through setup in the [Azure portal](https://portal.azure.com), where you can create a HDInsight cluster.
37
37
38
38
## Basics
39
39
@@ -136,9 +136,9 @@ For more information, see [Use external metadata stores in Azure HDInsight](./hd
136
136
137
137
#### SQL database for Hive
138
138
139
-
If you want to retain your Hive tables after you delete an HDInsight cluster, use a custom metastore. You can then attach the metastore to another HDInsight cluster.
139
+
If you want to retain your Hive tables after you delete a HDInsight cluster, use a custom metastore. You can then attach the metastore to another HDInsight cluster.
140
140
141
-
An HDInsight metastore that is created for one HDInsight cluster version can't be shared across different HDInsight cluster versions. For a list of HDInsight versions, see [Supported HDInsight versions](hdinsight-component-versioning.md#supported-hdinsight-versions).
141
+
A HDInsight metastore that is created for one HDInsight cluster version can't be shared across different HDInsight cluster versions. For a list of HDInsight versions, see [Supported HDInsight versions](hdinsight-component-versioning.md#supported-hdinsight-versions).
142
142
143
143
> [!IMPORTANT]
144
144
> The default metastore provides an Azure SQL Database with a **basic tier 5 DTU limit (not upgradeable)**! Suitable for basic testing purposes. For large or production workloads, we recommend migrating to an external metastore.
@@ -194,7 +194,7 @@ For more information, see [Managed identities in Azure HDInsight](./hdinsight-ma
194
194
195
195
:::image type="content" source="./media/hdinsight-hadoop-provision-linux-clusters/azure-portal-cluster-configuration-disk-attach.png" alt-text="HDInsight choose your node size.":::
196
196
197
-
You're billed for node usage for as long as the cluster exists. Billing starts when a cluster is created and stops when the cluster is deleted. Clusters can't be de-allocated or put on hold.
197
+
You're billed for node usage for as long as the cluster exists. Billing starts when a cluster is created and stops when the cluster is deleted. Clusters can't be deallocated or put on hold.
198
198
199
199
### Node configuration
200
200
@@ -208,7 +208,7 @@ Each cluster type has its own number of nodes, terminology for nodes, and defaul
208
208
209
209
For more information, see [Default node configuration and virtual machine sizes for clusters](hdinsight-supported-node-configuration.md) in "What are the Hadoop components and versions in HDInsight?"
210
210
211
-
The cost of HDInsight clusters is determined by the number of nodes and the virtual machines sizes for the nodes.
211
+
The cost of HDInsight clusters determined by the number of nodes and the virtual machines sizes for the nodes.
212
212
213
213
Different cluster types have different node types, numbers of nodes, and node sizes:
214
214
@@ -245,7 +245,7 @@ For more information, see [Sizes for virtual machines](/azure/virtual-machines/s
245
245
> The added disks are only configured for node manager local directories and **not for datanode directories**
246
246
247
247
248
-
HDInsight cluster comes with pre-defined disk space based on SKU. If you run some large applications, can lead to insufficient disk space, with disk full error - `LinkId=221672#ERROR_NOT_ENOUGH_DISK_SPACE` and job failures.
248
+
HDInsight cluster comes with predefined disk space based on SKU. If you run some large applications, can lead to insufficient disk space, with disk full error - `LinkId=221672#ERROR_NOT_ENOUGH_DISK_SPACE` and job failures.
249
249
250
250
More discs can be added to the cluster using the new feature **NodeManager**’s local directory. At the time of Hive and Spark cluster creation, the number of discs can be selected and added to the worker nodes. The selected disk, which will be of size 1TB each, would be part of **NodeManager**'s local directories.
0 commit comments