Skip to content

Commit 3b28a5e

Browse files
committed
freshness_c56
1 parent ae70188 commit 3b28a5e

File tree

1 file changed

+8
-8
lines changed

1 file changed

+8
-8
lines changed

articles/hdinsight/hdinsight-changing-configs-via-ambari.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -6,13 +6,13 @@ ms.author: hrasheed
66
ms.reviewer: jasonh
77
ms.service: hdinsight
88
ms.topic: conceptual
9-
ms.custom: hdinsightactive
10-
ms.date: 04/16/2020
9+
ms.custom: hdinsightactive,seoapr2020
10+
ms.date: 04/28/2020
1111
---
1212

1313
# Use Apache Ambari to optimize HDInsight cluster configurations
1414

15-
HDInsight provides [Apache Hadoop](./hadoop/apache-hadoop-introduction.md) clusters for large-scale data processing applications. Managing, monitoring, and optimizing these complex multi-node clusters can be challenging. [Apache Ambari](https://ambari.apache.org/) is a web interface to manage and monitor HDInsight Linux clusters. For Windows clusters, use the [Ambari REST API](hdinsight-hadoop-manage-ambari-rest-api.md).
15+
HDInsight provides Apache Hadoop clusters for large-scale data processing applications. Managing, monitoring, and optimizing these complex multi-node clusters can be challenging. Apache Ambari is a web interface to manage and monitor HDInsight Linux clusters. For Windows clusters, use the [Ambari REST API](hdinsight-hadoop-manage-ambari-rest-api.md).
1616

1717
For an introduction to using the Ambari Web UI, see [Manage HDInsight clusters by using the Apache Ambari Web UI](hdinsight-hadoop-manage-ambari.md)
1818

@@ -63,7 +63,7 @@ The following sections describe configuration options for optimizing overall Apa
6363

6464
### Set the Hive execution engine
6565

66-
Hive provides two execution engines: [Apache Hadoop MapReduce](https://hadoop.apache.org/docs/r1.2.1/mapred_tutorial.html) and [Apache TEZ](https://tez.apache.org/). Tez is faster than MapReduce. HDInsight Linux clusters have Tez as the default execution engine. To change the execution engine:
66+
Hive provides two execution engines: Apache Hadoop MapReduce and Apache TEZ. Tez is faster than MapReduce. HDInsight Linux clusters have Tez as the default execution engine. To change the execution engine:
6767

6868
1. In the Hive **Configs** tab, type **execution engine** in the filter box.
6969

@@ -94,7 +94,7 @@ These changes affect all Tez jobs across the server. To get an optimal result,
9494

9595
### Tune reducers
9696

97-
[Apache ORC](https://orc.apache.org/) and [Snappy](https://google.github.io/snappy/) both offer high performance. However, Hive may have too few reducers by default, causing bottlenecks.
97+
Apache ORC and Snappy both offer high performance. However, Hive may have too few reducers by default, causing bottlenecks.
9898

9999
For example, say you have an input data size of 50 GB. That data in ORC format with Snappy compression is 1 GB. Hive estimates the number of reducers needed as: (number of bytes input to mappers / `hive.exec.reducers.bytes.per.reducer`).
100100

@@ -281,7 +281,7 @@ Additional recommendations for optimizing the Hive execution engine:
281281

282282
## Apache Pig optimization
283283

284-
[Apache Pig](https://pig.apache.org/) properties can be modified from the Ambari web UI to tune Pig queries. Modifying Pig properties from Ambari directly modifies the Pig properties in the `/etc/pig/2.4.2.0-258.0/pig.properties` file.
284+
Apache Pig properties can be modified from the Ambari web UI to tune Pig queries. Modifying Pig properties from Ambari directly modifies the Pig properties in the `/etc/pig/2.4.2.0-258.0/pig.properties` file.
285285

286286
1. To modify Pig properties, navigate to the Pig **Configs** tab, and then expand the **Advanced pig-properties** pane.
287287

@@ -332,7 +332,7 @@ Pig generates temporary files during job execution. Compressing the temporary fi
332332

333333
* `pig.tmpfilecompression`: When true, enables temporary file compression. The default value is false.
334334

335-
* `pig.tmpfilecompression.codec`: The compression codec to use for compressing the temporary files. The recommended compression codecs are [LZO](https://www.oberhumer.com/opensource/lzo/) and Snappy for lower CPU use.
335+
* `pig.tmpfilecompression.codec`: The compression codec to use for compressing the temporary files. The recommended compression codecs are LZO and Snappy for lower CPU use.
336336

337337
### Enable split combining
338338

@@ -348,7 +348,7 @@ The number of reducers is calculated based on the parameter `pig.exec.reducers.b
348348

349349
## Apache HBase optimization with the Ambari web UI
350350

351-
[Apache HBase](https://hbase.apache.org/) configuration is modified from the **HBase Configs** tab. The following sections describe some of the important configuration settings that affect HBase performance.
351+
Apache HBase configuration is modified from the **HBase Configs** tab. The following sections describe some of the important configuration settings that affect HBase performance.
352352

353353
### Set HBASE_HEAPSIZE
354354

0 commit comments

Comments
 (0)