Skip to content

Commit 2bbb990

Browse files
authored
Merge pull request #97391 from dagiro/freshness80
freshness80
2 parents a717983 + ee18045 commit 2bbb990

File tree

2 files changed

+32
-44
lines changed

2 files changed

+32
-44
lines changed
Lines changed: 32 additions & 44 deletions
Original file line numberDiff line numberDiff line change
@@ -1,14 +1,13 @@
11
---
22
title: Hue with Hadoop on HDInsight Linux-based clusters - Azure
33
description: Learn how to install Hue on HDInsight clusters and use tunneling to route the requests to Hue. Use Hue to browse storage and run Hive or Pig.
4-
keywords: hue hadoop
54
author: hrasheed-msft
6-
ms.date: 12/11/2017
75
ms.author: hrasheed
86
ms.reviewer: jasonh
97
ms.service: hdinsight
10-
ms.custom: hdinsightactive,hdiseo17may2017
118
ms.topic: conceptual
9+
ms.custom: hdinsightactive,hdiseo17may2017
10+
ms.date: 11/28/2019
1211
---
1312

1413
# Install and use Hue on HDInsight Hadoop clusters
@@ -33,57 +32,46 @@ Hue is a set of Web applications used to interact with an Apache Hadoop cluster.
3332
3433
## Install Hue using Script Actions
3534

36-
The script to install Hue on a Linux-based HDInsight cluster is available at https://hdiconfigactions.blob.core.windows.net/linuxhueconfigactionv02/install-hue-uber-v02.sh. You can use this script to install Hue on clusters with either Azure Storage Blobs (WASB) or Azure Data Lake Storage as default storage.
37-
38-
This section provides instructions about how to use the script when provisioning the cluster using the Azure portal.
35+
Use the information in the table below for your Script Action. See [Customize HDInsight clusters with Script Actions](hdinsight-hadoop-customize-cluster-linux.md) for specific instructions on using Script Actions.
3936

4037
> [!NOTE]
41-
> Azure PowerShell, the Azure Classic CLI, the HDInsight .NET SDK, or Azure Resource Manager templates can also be used to apply script actions. You can also apply script actions to already running clusters. For more information, see [Customize HDInsight clusters with Script Actions](hdinsight-hadoop-customize-cluster-linux.md).
42-
43-
1. Start provisioning a cluster by using the steps in [Provision HDInsight clusters on Linux](hdinsight-hadoop-provision-linux-clusters.md), but do not complete provisioning.
44-
45-
> [!NOTE]
46-
> To install Hue on HDInsight clusters, the recommended headnode size is at least A4 (8 cores, 14 GB memory).
47-
48-
1. On the **Optional Configuration** blade, select **Script Actions**, and provide the information as shown below:
49-
50-
![Provide script action parameters for Hue](./media/hdinsight-hadoop-hue-linux/hdi-hue-script-action.png "Provide script action parameters for Hue")
38+
> To install Hue on HDInsight clusters, the recommended headnode size is at least A4 (8 cores, 14 GB memory).
5139
52-
* **NAME**: Enter a friendly name for the script action.
53-
* **SCRIPT URI**: https://hdiconfigactions.blob.core.windows.net/linuxhueconfigactionv02/install-hue-uber-v02.sh
54-
* **HEAD**: Check this option.
55-
* **WORKER**: Leave this blank.
56-
* **ZOOKEEPER**: Leave this blank.
57-
* **PARAMETERS**: Leave this blank.
58-
59-
1. At the bottom of the **Script Actions**, use the **Select** button to save the configuration. Finally, use the **Select** button at the bottom of the **Optional Configuration** blade to save the optional configuration information.
60-
61-
1. Continue provisioning the cluster as described in [Provision HDInsight clusters on Linux](hdinsight-hadoop-provision-linux-clusters.md).
40+
|Property |Value |
41+
|---|---|
42+
|Script type:|- Custom|
43+
|Name|Install Hue|
44+
|Bash script URI|`https://hdiconfigactions.blob.core.windows.net/linuxhueconfigactionv02/install-hue-uber-v02.sh`|
45+
|Node type(s):|Head|
6246

6347
## Use Hue with HDInsight clusters
6448

65-
SSH Tunneling is the only way to access Hue on the cluster once it is running. Tunneling via SSH allows the traffic to go directly to the headnode of the cluster where Hue is running. After the cluster has finished provisioning, use the following steps to use Hue on an HDInsight Linux cluster.
49+
SSH Tunneling is the only way to access Hue on the cluster once it is running. Tunneling via SSH allows the traffic to go directly to the headnode of the cluster where Hue is running. After the cluster has finished provisioning, use the following steps to use Hue on an HDInsight cluster.
6650

6751
> [!NOTE]
6852
> We recommend using Firefox web browser to follow the instructions below.
6953
7054
1. Use the information in [Use SSH Tunneling to access Apache Ambari web UI, ResourceManager, JobHistory, NameNode, Oozie, and other web UI's](hdinsight-linux-ambari-ssh-tunnel.md) to create an SSH tunnel from your client system to the HDInsight cluster, and then configure your Web browser to use the SSH tunnel as a proxy.
7155

72-
2. Once you have created an SSH tunnel and configured your browser to proxy traffic through it, you must find the host name of the primary head node. You can do this by connecting to the cluster using SSH on port 22. For example, `ssh [email protected]` where **USERNAME** is your SSH user name and **CLUSTERNAME** is the name of your cluster.
56+
1. Use [ssh command](./hdinsight-hadoop-linux-use-ssh-unix.md) to connect to your cluster. Edit the command below by replacing CLUSTERNAME with the name of your cluster, and then enter the command:
7357

74-
For more information, see [Use SSH with HDInsight](hdinsight-hadoop-linux-use-ssh-unix.md).
58+
```cmd
59+
60+
```
7561
76-
3. Once connected, use the following command to obtain the fully qualified domain name of the primary headnode:
62+
1. Once connected, use the following command to obtain the fully qualified domain name of the primary headnode:
7763
78-
hostname -f
64+
```bash
65+
hostname -f
66+
```
7967
8068
This will return a name similar to the following:
8169
8270
hn0-myhdi-nfebtpfdv1nubcidphpap2eq2b.ex.internal.cloudapp.net
8371
8472
This is the hostname of the primary headnode where the Hue website is located.
8573
86-
4. Use the browser to open the Hue portal at http:\//HOSTNAME:8888. Replace HOSTNAME with the name you obtained in the previous step.
74+
1. Use the browser to open the Hue portal at `http://HOSTNAME:8888`. Replace HOSTNAME with the name you obtained in the previous step.
8775
8876
> [!NOTE]
8977
> When you log in for the first time, you will be prompted to create an account to log in to the Hue portal. The credentials you specify here will be limited to the portal and are not related to the admin or SSH user credentials you specified while provision the cluster.
@@ -92,9 +80,10 @@ SSH Tunneling is the only way to access Hue on the cluster once it is running. T
9280
9381
### Run a Hive query
9482
95-
1. From the Hue portal, click **Query Editors**, and then click **Hive** to open the Hive editor.
83+
1. From the Hue portal, select **Query Editors**, and then select **Hive** to open the Hive editor.
9684
9785
![HDInsight hue portal use hive editor](./media/hdinsight-hadoop-hue-linux/hdinsight-hue-portal-use-hive.png "Use Hive")
86+
9887
2. On the **Assist** tab, under **Database**, you should see **hivesampletable**. This is a sample table that is shipped with all Hadoop clusters on HDInsight. Enter a sample query in the right pane and see the output on the **Results** tab in the pane below, as shown in the screen capture.
9988
10089
![HDInsight hue portal hive query](./media/hdinsight-hadoop-hue-linux/hdinsight-hue-portal-hive-query.png "Run Hive query")
@@ -103,37 +92,36 @@ SSH Tunneling is the only way to access Hue on the cluster once it is running. T
10392
10493
### Browse the cluster storage
10594
106-
1. From the Hue portal, click **File Browser** in the top-right corner of the menu bar.
107-
2. By default the file browser opens at the **/user/myuser** directory. Click the forward slash right before the user directory in the path to go to the root of the Azure storage container associated with the cluster.
95+
1. From the Hue portal, select **File Browser** in the top-right corner of the menu bar.
96+
2. By default the file browser opens at the **/user/myuser** directory. Select the forward slash right before the user directory in the path to go to the root of the Azure storage container associated with the cluster.
10897
10998
![HDInsight hue portal file browser](./media/hdinsight-hadoop-hue-linux/hdinsight-hue-portal-file-browser.png "Use file browser")
11099
111100
3. Right-click on a file or folder to see the available operations. Use the **Upload** button in the right corner to upload files to the current directory. Use the **New** button to create new files or directories.
112101
113102
> [!NOTE]
114-
> The Hue file browser can only show the contents of the default container associated with the HDInsight cluster. Any additional storage accounts/containers that you might have associated with the cluster will not be accessible using the file browser. However, the additional containers associated with the cluster will always be accessible for the Hive jobs. For example, if you enter the command `dfs -ls wasb://[email protected]` in the Hive editor, you can see the contents of additional containers as well. In this command, **newcontainer** is not the default container associated with a cluster.
103+
> The Hue file browser can only show the contents of the default container associated with the HDInsight cluster. Any additional storage accounts/containers that you might have associated with the cluster will not be accessible using the file browser. However, the additional containers associated with the cluster will always be accessible for the Hive jobs. For example, if you enter the command `dfs -ls wasbs://[email protected]` in the Hive editor, you can see the contents of additional containers as well. In this command, **newcontainer** is not the default container associated with a cluster.
115104
116105
## Important considerations
117106
118107
1. The script used to install Hue installs it only on the primary headnode of the cluster.
119108
120-
2. During installation, multiple Hadoop services (HDFS, YARN, MR2, Oozie) are restarted for updating the configuration. After the script finishes installing Hue, it might take some time for other Hadoop services to start up. This might affect Hue's performance initially. Once all services start up, Hue will be fully functional.
121-
3. Hue does not understand Apache Tez jobs, which is the current default for Hive. If you want to use MapReduce as the Hive execution engine, update the script to use the following command in your script:
109+
1. During installation, multiple Hadoop services (HDFS, YARN, MR2, Oozie) are restarted for updating the configuration. After the script finishes installing Hue, it might take some time for other Hadoop services to start up. This might affect Hue's performance initially. Once all services start up, Hue will be fully functional.
110+
111+
1. Hue does not understand Apache Tez jobs, which is the current default for Hive. If you want to use MapReduce as the Hive execution engine, update the script to use the following command in your script:
122112
123113
set hive.execution.engine=mr;
124114
125-
4. With Linux clusters, you can have a scenario where your services are running on the primary headnode while the Resource Manager could be running on the secondary. Such a scenario might result in errors (shown below) when using Hue to view details of RUNNING jobs on the cluster. However, you can view the job details when the job has completed.
115+
1. With Linux clusters, you can have a scenario where your services are running on the primary headnode while the Resource Manager could be running on the secondary. Such a scenario might result in errors (shown below) when using Hue to view details of RUNNING jobs on the cluster. However, you can view the job details when the job has completed.
126116
127117
![Hue portal error sample message](./media/hdinsight-hadoop-hue-linux/hdinsight-hue-portal-error.png "Hue portal error")
128118
129119
This is due to a known issue. As a workaround, modify Ambari so that the active Resource Manager also runs on the primary headnode.
130-
5. Hue understands WebHDFS while HDInsight clusters use Azure Storage using `wasb://`. So, the custom script used with script action installs WebWasb, which is a WebHDFS-compatible service for talking to WASB. So, even though the Hue portal says HDFS in places (like when you move your mouse over the **File Browser**), it should be interpreted as WASB.
120+
121+
1. Hue understands WebHDFS while HDInsight clusters use Azure Storage using `wasbs://`. So, the custom script used with script action installs WebWasb, which is a WebHDFS-compatible service for talking to WASB. So, even though the Hue portal says HDFS in places (like when you move your mouse over the **File Browser**), it should be interpreted as WASB.
131122
132123
## Next steps
133124
134125
* [Install Apache Giraph on HDInsight clusters](hdinsight-hadoop-giraph-install-linux.md). Use cluster customization to install Giraph on HDInsight Hadoop clusters. Giraph allows you to perform graph processing using Hadoop, and it can be used with Azure HDInsight.
135-
* [Install R on HDInsight clusters](hdinsight-hadoop-r-scripts-linux.md). Use cluster customization to install R on HDInsight Hadoop clusters. R is an open-source language and environment for statistical computing. It provides hundreds of built-in statistical functions and its own programming language that combines aspects of functional and object-oriented programming. It also provides extensive graphical capabilities.
136126
137-
[powershell-install-configure]: install-configure-powershell-linux.md
138-
[hdinsight-provision]: hdinsight-provision-clusters-linux.md
139-
[hdinsight-cluster-customize]: hdinsight-hadoop-customize-cluster-linux.md
127+
* [Install R on HDInsight clusters](hdinsight-hadoop-r-scripts-linux.md). Use cluster customization to install R on HDInsight Hadoop clusters. R is an open-source language and environment for statistical computing. It provides hundreds of built-in statistical functions and its own programming language that combines aspects of functional and object-oriented programming. It also provides extensive graphical capabilities.
Binary file not shown.

0 commit comments

Comments
 (0)