You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/hdinsight/hbase/apache-hbase-tutorial-get-started-linux.md
+11-11Lines changed: 11 additions & 11 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -4,7 +4,7 @@ description: Follow this Apache HBase tutorial to start using hadoop on HDInsigh
4
4
ms.service: azure-hdinsight
5
5
ms.topic: tutorial
6
6
ms.custom: hdinsightactive, linux-related-content
7
-
ms.date: 05/10/2024
7
+
ms.date: 12/23/2024
8
8
---
9
9
10
10
# Tutorial: Use Apache HBase in Azure HDInsight
@@ -24,13 +24,13 @@ In this tutorial, you learn how to:
24
24
25
25
* An SSH client. For more information, see [Connect to HDInsight (Apache Hadoop) using SSH](../hdinsight-hadoop-linux-use-ssh-unix.md).
26
26
27
-
* Bash. The examples in this article use the Bash shell on Windows 10 for the curl commands. See [Windows Subsystem for Linux Installation Guide for Windows 10](/windows/wsl/install-win10) for installation steps. Other [Unix shells](https://www.gnu.org/software/bash/)will work as well. The curl examples, with some slight modifications, can work on a Windows Command prompt. Or you can use the Windows PowerShell cmdlet [Invoke-RestMethod](/powershell/module/microsoft.powershell.utility/invoke-restmethod).
27
+
* Bash. The examples in this article use the Bash shell on Windows 10 for the curl commands. See [Windows Subsystem for Linux Installation Guide for Windows 10](/windows/wsl/install-win10) for installation steps. Other [Unix shells](https://www.gnu.org/software/bash/) work as well. The curl examples, with some slight modifications, can work on a Windows Command prompt. Or you can use the Windows PowerShell cmdlet [Invoke-RestMethod](/powershell/module/microsoft.powershell.utility/invoke-restmethod).
28
28
29
29
## Create Apache HBase cluster
30
30
31
-
The following procedure uses an Azure Resource Manager template to create an HBase cluster. The template also creates the dependent default Azure Storage account. To understand the parameters used in the procedure and other cluster creation methods, see [Create Linux-based Hadoop clusters in HDInsight](../hdinsight-hadoop-provision-linux-clusters.md).
31
+
The following procedure uses an Azure Resource Manager template to create a HBase cluster. The template also creates the dependent default Azure Storage account. To understand the parameters used in the procedure and other cluster creation methods, see [Create Linux-based Hadoop clusters in HDInsight](../hdinsight-hadoop-provision-linux-clusters.md).
32
32
33
-
1. Select the following image to open the template in the Azure portal. The template is located in [Azure quickstart templates](https://azure.microsoft.com/resources/templates/).
33
+
1. Select the following image to open the template in the Azure portal. The template is located in [Azure Quickstart templates](https://azure.microsoft.com/resources/templates/).
34
34
35
35
<ahref="https://portal.azure.com/#create/Microsoft.Template/uri/https%3A%2F%2Fraw.githubusercontent.com%2FAzure%2Fazure-quickstart-templates%2Fmaster%2Fquickstarts%2Fmicrosoft.hdinsight%2Fhdinsight-hbase-linux%2Fazuredeploy.json"target="_blank"><imgsrc="./media/apache-hbase-tutorial-get-started-linux/hdi-deploy-to-azure1.png"alt="Deploy to Azure button for new cluster"></a>
36
36
@@ -51,7 +51,7 @@ The following procedure uses an Azure Resource Manager template to create an HBa
51
51
52
52
3. Select **I agree to the terms and conditions stated above**, and then select **Purchase**. It takes about 20 minutes to create a cluster.
53
53
54
-
After an HBase cluster is deleted, you can create another HBase cluster by using the same default blob container. The new cluster picks up the HBase tables you created in the original cluster. To avoid inconsistencies, we recommend that you disable the HBase tables before you delete the cluster.
54
+
After a HBase cluster is deleted, you can create another HBase cluster by using the same default blob container. The new cluster picks up the HBase tables you created in the original cluster. To avoid inconsistencies, we recommend that you disable the HBase tables before you delete the cluster.
55
55
56
56
## Create tables and insert data
57
57
@@ -67,7 +67,7 @@ In HBase (an implementation of [Cloud BigTable](https://cloud.google.com/bigtabl
67
67
68
68
**To use the HBase shell**
69
69
70
-
1. Use `ssh` command to connect to your HBase cluster. Edit the command below by replacing `CLUSTERNAME` with the name of your cluster, and then enter the command:
70
+
1. Use `ssh` command to connect to your HBase cluster. Edit the following command by replacing `CLUSTERNAME` with the name of your cluster, and then enter the command:
@@ -79,7 +79,7 @@ In HBase (an implementation of [Cloud BigTable](https://cloud.google.com/bigtabl
79
79
hbase shell
80
80
```
81
81
82
-
1. Use `create` command to create an HBase table with two-column families. The table and column names are case-sensitive. Enter the following command:
82
+
1. Use `create` command to create a HBase table with two-column families. The table and column names are case-sensitive. Enter the following command:
83
83
84
84
```hbaseshell
85
85
create 'Contacts', 'Personal', 'Office'
@@ -204,15 +204,15 @@ You can query data in HBase tables by using [Apache Hive](https://hive.apache.or
204
204
The Hive query to access HBase data need not be executed from the HBase cluster. Any cluster that comes with Hive (including Spark, Hadoop, HBase, or Interactive Query) can be used to query HBase data, provided the following steps are completed:
205
205
206
206
1. Both clusters must be attached to the same Virtual Network and Subnet
207
-
2. Copy `/usr/hdp/$(hdp-select --version)/hbase/conf/hbase-site.xml` from the HBase cluster headnodes to the Hive cluster headnodes and workernodes.
207
+
2. Copy `/usr/hdp/$(hdp-select --version)/hbase/conf/hbase-site.xml` from the HBase cluster headnodes to the Hive cluster headnodes and worker nodes.
208
208
209
209
### Secure Clusters
210
210
211
211
HBase data can also be queried from Hive using ESP-enabled HBase:
212
212
213
213
1. When following a multi-cluster pattern, both clusters must be ESP-enabled.
214
214
2. To allow Hive to query HBase data, make sure that the `hive` user is granted permissions to access the HBase data via the Hbase Apache Ranger plugin
215
-
3. When using separate, ESP-enabled clusters, the contents of `/etc/hosts` from the HBase cluster headnodes must be appended to `/etc/hosts` of the Hive cluster headnodes and workernodes.
215
+
3. When you use separate, ESP-enabled clusters, the contents of `/etc/hosts` from the HBase cluster headnodes must be appended to `/etc/hosts` of the Hive cluster headnodes and worker nodes.
216
216
> [!NOTE]
217
217
> After scaling either clusters, `/etc/hosts` must be appended again
218
218
@@ -245,7 +245,7 @@ The HBase REST API is secured via [basic authentication](https://en.wikipedia.or
245
245
fi
246
246
```
247
247
248
-
1. Set environment variable for ease of use. Edit the commands below by replacing `MYPASSWORD` with the cluster login password. Replace `MYCLUSTERNAME` with the name of your HBase cluster. Then enter the commands.
248
+
1. Set environment variable for ease of use. Edit the following commands by replacing `MYPASSWORD` with the cluster login password. Replace `MYCLUSTERNAME` with the name of your HBase cluster. Then enter the commands.
249
249
250
250
```bash
251
251
export PASSWORD='MYPASSWORD'
@@ -307,7 +307,7 @@ For more information about HBase Rest, see [Apache HBase Reference Guide](https:
307
307
> [!NOTE]
308
308
> Thrift is not supported by HBase in HDInsight.
309
309
>
310
-
> When using Curl or any other REST communication with WebHCat, you must authenticate the requests by providing the user name and password for the HDInsight cluster administrator. You must also use the cluster name as part of the Uniform Resource Identifier (URI) used to send the requests to the server:
310
+
> When you use Curl or any other REST communication with WebHCat, you must authenticate the requests by providing the user name and password for the HDInsight cluster administrator. You must also use the cluster name as part of the Uniform Resource Identifier (URI) used to send the requests to the server:
0 commit comments