Skip to content

Commit 3508ba7

Browse files
authored
Merge pull request #276034 from sreekzz/freshness-date-change
MS Freshness Date Change
2 parents cd235e5 + a477614 commit 3508ba7

15 files changed

+91
-91
lines changed

articles/hdinsight/domain-joined/apache-domain-joined-run-hive.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ title: Apache Hive policies in Apache Ranger - Azure HDInsight
33
description: Learn how to configure Apache Ranger policies for Hive in an Azure HDInsight service with Enterprise Security Package.
44
ms.service: hdinsight
55
ms.topic: conceptual
6-
ms.date: 04/11/2023
6+
ms.date: 05/22/2024
77
---
88

99
# Configure Apache Hive policies in HDInsight with Enterprise Security Package
@@ -103,7 +103,7 @@ In the last section, you configured two policies: `hiveuser1` has the select per
103103

104104
1. For the first use, an **ODBC driver** dialog opens. Select **Windows** from the left menu. Then select **Connect** to open the **Navigator** window.
105105

106-
1. Wait for the **Select Database and Table** dialog to open. This step can take a few seconds.
106+
1. Wait for the `Select Database and Table` dialog to open. This step can take a few seconds.
107107

108108
1. Select **hivesampletable** > **Next**.
109109

articles/hdinsight/hadoop/apache-hadoop-on-premises-migration-best-practices-security-devops.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ description: Learn security and DevOps best practices for migrating on-premises
44
ms.service: hdinsight
55
ms.topic: how-to
66
ms.custom: hdinsightactive
7-
ms.date: 04/26/2023
7+
ms.date: 05/22/2024
88
---
99

1010
# Migrate on-premises Apache Hadoop clusters to Azure HDInsight - security and DevOps best practices

articles/hdinsight/hadoop/apache-hadoop-on-premises-migration-best-practices-storage.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ description: Learn storage best practices for migrating on-premises Hadoop clust
44
ms.service: hdinsight
55
ms.topic: how-to
66
ms.custom: hdinsightactive
7-
ms.date: 01/04/2024
7+
ms.date: 05/22/2024
88
---
99

1010
# Migrate on-premises Apache Hadoop clusters to Azure HDInsight

articles/hdinsight/hadoop/apache-hadoop-use-mapreduce-dotnet-sdk.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -4,14 +4,14 @@ description: Learn how to submit MapReduce jobs to Azure HDInsight Apache Hadoop
44
ms.service: hdinsight
55
ms.topic: how-to
66
ms.custom: hdinsightactive, devx-track-csharp, devx-track-dotnet
7-
ms.date: 04/24/2023
7+
ms.date: 05/22/2024
88
---
99

1010
# Run MapReduce jobs using HDInsight .NET SDK
1111

1212
[!INCLUDE [mapreduce-selector](../includes/hdinsight-selector-use-mapreduce.md)]
1313

14-
Learn how to submit MapReduce jobs using HDInsight .NET SDK. HDInsight clusters come with a jar file with some MapReduce samples. The jar file is `/example/jars/hadoop-mapreduce-examples.jar`. One of the samples is **wordcount**. You develop a C# console application to submit a wordcount job. The job reads the `/example/data/gutenberg/davinci.txt` file, and outputs the results to `/example/data/davinciwordcount`. If you want to rerun the application, you must clean up the output folder.
14+
Learn how to submit MapReduce jobs using HDInsight .NET SDK. HDInsight clusters come with a jar file with some MapReduce samples. The jar file is `/example/jars/hadoop-mapreduce-examples.jar`. One of the samples is **wordcount**. You develop a C# console application to submit a wordcount job. The job reads the `/example/data/gutenberg/davinci.txt` file, and outputs the results to `/example/data/davinciwordcount`. If you want to rerun the application, you must clean up the output folder.
1515

1616
> [!NOTE]
1717
> The steps in this article must be performed from a Windows client. For information on using a Linux, OS X, or Unix client to work with Hive, use the tab selector shown on the top of the article.
@@ -161,7 +161,7 @@ When the job completes successfully, the application prints the content of the o
161161
162162
## Next steps
163163
164-
In this article, you have learned several ways to create an HDInsight cluster. To learn more, see the following articles:
164+
In this article, you learned several ways to create an HDInsight cluster. To learn more, see the following articles:
165165
166166
* For submitting a Hive job, see [Run Apache Hive queries using HDInsight .NET SDK](apache-hadoop-use-hive-dotnet-sdk.md).
167167
* For creating HDInsight clusters, see [Create Linux-based Apache Hadoop clusters in HDInsight](../hdinsight-hadoop-provision-linux-clusters.md).

articles/hdinsight/hdinsight-autoscale-clusters.md

Lines changed: 23 additions & 23 deletions
Large diffs are not rendered by default.

articles/hdinsight/hdinsight-hadoop-manage-ambari-rest-api.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,10 @@
11
---
22
title: Monitor and manage Hadoop with Ambari REST API - Azure HDInsight
3-
description: Learn how to use Ambari to monitor and manage Hadoop clusters in Azure HDInsight. In this document, you'll learn how to use the Ambari REST API included with HDInsight clusters.
3+
description: Learn how to use Ambari to monitor and manage Hadoop clusters in Azure HDInsight. In this document, you learn how to use the Ambari REST API included with HDInsight clusters.
44
ms.service: hdinsight
55
ms.topic: how-to
66
ms.custom: hdinsightactive
7-
ms.date: 07/20/2023
7+
ms.date: 05/22/2024
88
---
99

1010
# Manage HDInsight clusters by using the Apache Ambari REST API
@@ -21,7 +21,7 @@ Apache Ambari simplifies the management and monitoring of Hadoop clusters by pro
2121

2222
* A Hadoop cluster on HDInsight. See [Get Started with HDInsight on Linux](hadoop/apache-hadoop-linux-tutorial-get-started.md).
2323

24-
* Bash on Ubuntu on Windows 10. The examples in this article use the Bash shell on Windows 10. See [Windows Subsystem for Linux Installation Guide for Windows 10](/windows/wsl/install-win10) for installation steps. Other [Unix shells](https://www.gnu.org/software/bash/) will work as well. The examples, with some slight modifications, can work on a Windows Command prompt. Or you can use Windows PowerShell.
24+
* Bash on Ubuntu on Windows 10. The examples in this article use the Bash shell on Windows 10. See [Windows Subsystem for Linux Installation Guide for Windows 10](/windows/wsl/install-win10) for installation steps. Other [Unix shells](https://www.gnu.org/software/bash/) work as well. The examples, with some slight modifications, can work on a Windows Command prompt. Or you can use Windows PowerShell.
2525

2626
* jq, a command-line JSON processor. See [https://stedolan.github.io/jq/](https://stedolan.github.io/jq/).
2727

@@ -41,10 +41,10 @@ For Enterprise Security Package clusters, instead of `admin`, use a fully qualif
4141

4242
### Setup (Preserve credentials)
4343

44-
Preserve your credentials to avoid reentering them for each example. The cluster name will be preserved in a separate step.
44+
Preserve your credentials to avoid reentering them for each example. The cluster name is preserved in a separate step.
4545

4646
**A. Bash**
47-
Edit the script below by replacing `PASSWORD` with your actual password. Then enter the command.
47+
Edit the script by replacing `PASSWORD` with your actual password. Then enter the command.
4848

4949
```bash
5050
export password='PASSWORD'
@@ -58,9 +58,9 @@ $creds = Get-Credential -UserName "admin" -Message "Enter the HDInsight login"
5858

5959
### Identify correctly cased cluster name
6060

61-
The actual casing of the cluster name may be different than you expect. The steps here will show the actual casing, and then store it in a variable for all later examples.
61+
The actual casing of the cluster name may be different than you expect. The following steps show the actual casing, and then store it in a variable for all later examples.
6262

63-
Edit the scripts below to replace `CLUSTERNAME` with your cluster name. Then enter the command. (The cluster name for the FQDN isn't case-sensitive.)
63+
Edit the scripts to replace `CLUSTERNAME` with your cluster name. Then enter the command. (The cluster name for the FQDN isn't case-sensitive.)
6464

6565
```bash
6666
export clusterName=$(curl -u admin:$password -sS -G "https://CLUSTERNAME.azurehdinsight.net/api/v1/clusters" | jq -r '.items[].Clusters.cluster_name')
@@ -297,7 +297,7 @@ This example returns a JSON document containing the current configuration for th
297297
### Update configuration
298298
299299
1. Create `newconfig.json`.
300-
Modify, and then enter the commands below:
300+
Modify, and then enter the commands as follows:
301301
302302
* Replace `livy2-conf` with the new component.
303303
* Replace `INITIAL` with actual value retrieved for `tag` from [Get all configurations](#get-all-configurations).

articles/hdinsight/hdinsight-hadoop-script-actions-linux.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ description: Learn how to use Bash scripts to customize HDInsight clusters. Scri
44
ms.service: hdinsight
55
ms.custom: linux-related-content
66
ms.topic: how-to
7-
ms.date: 04/26/2023
7+
ms.date: 05/22/2024
88
---
99

1010
# Script action development with HDInsight
@@ -68,7 +68,7 @@ fi
6868

6969
### <a name="bps10"></a> Target the operating system version
7070

71-
HDInsight is based on the Ubuntu Linux distribution. Different versions of HDInsight rely on different versions of Ubuntu, which may change how your script behaves. For example, HDInsight 3.4 and earlier are based on Ubuntu versions that use Upstart. Versions 3.5 and greater are based on Ubuntu 16.04, which uses Systemd. Systemd and Upstart rely on different commands, so your script should be written to work with both.
71+
HDInsight is based on the Ubuntu Linux distribution. Different versions of HDInsight rely on different versions of Ubuntu, which may change how your script behaves. For example, HDInsight 3.4 and earlier are based on Ubuntu versions that use Upstart. Versions 3.5 and greater are based on Ubuntu 16.04, which uses `Systemd`. `Systemd` and Upstart rely on different commands, so your script should be written to work with both.
7272

7373
Another important difference between HDInsight 3.4 and 3.5 is that `JAVA_HOME` now points to Java 8. The following code demonstrates how to determine if the script is running on Ubuntu 14 or 16:
7474

@@ -105,7 +105,7 @@ You can find the full script that contains these snippets at https://hdiconfigac
105105

106106
For the version of Ubuntu that is used by HDInsight, see the [HDInsight component version](hdinsight-component-versioning.md) document.
107107

108-
To understand the differences between Systemd and Upstart, see [Systemd for Upstart users](https://wiki.ubuntu.com/SystemdForUpstartUsers).
108+
To understand the differences between `Systemd` and Upstart, see [`Systemd` for Upstart users](https://wiki.ubuntu.com/SystemdForUpstartUsers).
109109

110110
### <a name="bPS2"></a>Provide stable links to script resources
111111

@@ -126,7 +126,7 @@ To reduce the time it takes to run the script, avoid operations that compile res
126126

127127
Scripts must be idempotent. If the script runs multiple times, it should return the cluster to the same state every time.
128128

129-
For example, a script that modifies configuration files shouldn't add duplicate entries if ran multiple times.
129+
If the script runs multiple times, the script that modifies configuration files shouldn't add duplicate entries.
130130

131131
### <a name="bPS5"></a>Ensure high availability of the cluster architecture
132132

@@ -181,7 +181,7 @@ line 1: #!/usr/bin/env: No such file or directory
181181

182182
### <a name="bps9"></a> Use retry logic to recover from transient errors
183183

184-
When downloading files, installing packages using apt-get, or other actions that transmit data over the internet, the action may fail because of transient networking errors. For example, the remote resource you're communicating with may be in the process of failing over to a backup node.
184+
When you download files, installing packages using apt-get, or other actions that transmit data over the internet, the action may fail because of transient networking errors. For example, the remote resource you're communicating with may be in the process of failing over to a backup node.
185185

186186
To make your script resilient to transient errors, you can implement retry logic. The following function demonstrates how to implement retry logic. It retries the operation three times before failing.
187187

articles/hdinsight/hdinsight-hadoop-use-blob-storage.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ title: Query data from HDFS-compatible Azure storage - Azure HDInsight
33
description: Learn how to query data from Azure storage and Azure Data Lake Storage to store results of your analysis.
44
ms.service: hdinsight
55
ms.topic: how-to
6-
ms.date: 04/24/2023
6+
ms.date: 05/22/2024
77
---
88

99
# Use Azure storage with Azure HDInsight clusters

articles/hdinsight/hdinsight-phoenix-in-hdinsight.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ description: Overview of Apache Phoenix
44
ms.service: hdinsight
55
ms.topic: how-to
66
ms.custom: hdinsightactive
7-
ms.date: 04/26/2023
7+
ms.date: 05/22/2024
88
---
99

1010
# Apache Phoenix in Azure HDInsight
@@ -21,7 +21,7 @@ Apache Phoenix adds several performance enhancements and features to HBase que
2121

2222
### Secondary indexes
2323

24-
HBase has a single index that is lexicographically sorted on the primary row key. These records can only be accessed through the row key. Accessing records through any column other than the row key requires scanning all of the data while applying the required filter. In a secondary index, the columns or expressions that are indexed form an alternate row key, allowing lookups and range scans on that index.
24+
HBase has a single index that is lexicographically sorted on the primary row key. These records can only be accessed through the row key. Accessing records through any column other than the row key requires scanning all of the data while applying the required filter. In a secondary index, the columns or expressions that are indexed from an alternate row key, allowing lookups and range scans on that index.
2525

2626
Create a secondary index with the `CREATE INDEX` command:
2727

@@ -37,7 +37,7 @@ Phoenix views provide a way to overcome an HBase limitation, where performance
3737

3838
Creating a Phoenix view is similar to using standard SQL view syntax. One difference is that you can define columns for your view, in addition to the columns inherited from its base table. You can also add new `KeyValue` columns.
3939

40-
For example, here is a physical table named `product_metrics` with the following definition:
40+
For example, here's a physical table named `product_metrics` with the following definition:
4141

4242
```sql
4343
CREATE TABLE product_metrics (
@@ -48,7 +48,7 @@ CREATE TABLE product_metrics (
4848
CONSTRAINT pk PRIMARY KEY (metric_type, created_by, created_date, metric_id));
4949
```
5050

51-
Define a view over this table, with additional columns:
51+
Define a view over this table, with more columns:
5252

5353
```sql
5454
CREATE VIEW mobile_product_metrics (carrier VARCHAR, dropped_calls BIGINT) AS
@@ -60,15 +60,15 @@ To add more columns later, use the `ALTER VIEW` statement.
6060

6161
### Skip scan
6262

63-
Skip scan uses one or more columns of a composite index to find distinct values. Unlike a range scan, skip scan implements intra-row scanning, yielding [improved performance](https://phoenix.apache.org/performance.html#Skip-Scan). While scanning, the first matched value is skipped along with the index until the next value is found.
63+
Skip scan uses one or more columns of a composite index to find distinct values. Unlike a range scan, skip scan implements intra-row scanning, yielding [improved performance](https://phoenix.apache.org/performance.html#Skip-Scan). When you scan, the first matched value is skipped along with the index until the next value is found.
6464

6565
A skip scan uses the `SEEK_NEXT_USING_HINT` enumeration of the HBase filter. Using `SEEK_NEXT_USING_HINT`, the skip scan keeps track of which set of keys, or ranges of keys, are being searched for in each column. The skip scan then takes a key that was passed to it during filter evaluation, and determines whether it's one of the combinations. If not, the skip scan evaluates the next highest key to jump to.
6666

6767
### Transactions
6868

6969
While HBase provides row-level transactions, Phoenix integrates with [Tephra](https://tephra.apache.org/) to add cross-row and cross-table transaction support with full [ACID](https://en.wikipedia.org/wiki/ACID) semantics.
7070

71-
As with traditional SQL transactions, transactions provided through the Phoenix transaction manager allow you to ensure an atomic unit of data is successfully upserted, rolling back the transaction if the upsert operation fails on any transaction-enabled table.
71+
As with traditional SQL transactions, transactions provided through the Phoenix transaction manager allow you to ensure an atomic unit of data is successfully upserted, rolling back the transaction if the upserted operation fails on any transaction-enabled table.
7272

7373
To enable Phoenix transactions, see the [Apache Phoenix transaction documentation](https://phoenix.apache.org/transactions.html).
7474

0 commit comments

Comments
 (0)