Skip to content

Commit 2efb2a8

Browse files
authored
Update apache-hive-warehouse-connector.md
1 parent a9b1663 commit 2efb2a8

File tree

1 file changed

+9
-9
lines changed

1 file changed

+9
-9
lines changed

articles/hdinsight/interactive-query/apache-hive-warehouse-connector.md

Lines changed: 9 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -20,8 +20,8 @@ Apache Hive offers support for database transactions that are Atomic, Consistent
2020
Apache Spark has a Structured Streaming API that gives streaming capabilities not available in Apache Hive. Beginning with HDInsight 4.0, Apache Spark 2.3.1 & above, and Apache Hive 3.1.0 have separate metastore catalogs, which make interoperability difficult.
2121

2222
The Hive Warehouse Connector (HWC) makes it easier to use Spark and Hive together. The HWC library loads data from LLAP daemons to Spark executors in parallel. This process makes it more efficient and adaptable than a standard JDBC connection from Spark to Hive. This brings out two different execution modes for HWC:
23-
> - Hive JDBC mode via HiveServer2
24-
> - Hive LLAP mode using LLAP daemons **[Recommended]**
23+
- Hive JDBC mode via HiveServer2
24+
- Hive LLAP mode using LLAP daemons **[Recommended]**
2525

2626
By default, HWC is configured to use Hive LLAP daemons.
2727
For executing Hive queries (both read and write) using the above modes with their respective APIs, see [HWC APIs](./hive-warehouse-connector-apis.md).
@@ -99,7 +99,7 @@ value. The value may be similar to: `thrift://iqgiro.rekufuk2y2cezcbowjkbwfnyvd.
9999

100100
1. Save changes and restart all affected components.
101101

102-
#### Additonal configurations for Spark and Hive
102+
#### Additional configurations for Spark and Hive
103103

104104
The following configuration needs to be done for **all** head and worker nodes of your Spark and Hive clusters.
105105

@@ -129,7 +129,7 @@ Apart from the configurations mentioned in the previous section, add the followi
129129
130130
* From a web browser, navigate to `https://CLUSTERNAME.azurehdinsight.net/#/main/services/HIVE/summary` where CLUSTERNAME is the name of your Interactive Query cluster. Click on **HiveServer2 Interactive**. You'll see the Fully Qualified Domain Name (FQDN) of the head node on which LLAP is running as shown in the screenshot. Replace `<llap-headnode>` with this value.
131131
132-
:::image type="content" source="./media/apache-hive-warehouse-connector/head-node-hive-server-interactive.png" alt-text="hive warehouse connector Head Node." border="true":::
132+
:::image type="content" source="./media/apache-hive-warehouse-connector/head-node-hive-server-interactive.png" alt-text="Screenshot of hive warehouse connector Head Node." border="true":::
133133
134134
* Use [ssh command](../hdinsight-hadoop-linux-use-ssh-unix.md) to connect to your Interactive Query cluster. Look for `default_realm` parameter in the `/etc/krb5.conf` file. Replace `<AAD-DOMAIN>` with this value as an uppercase string, otherwise the credential won't be found.
135135
@@ -256,21 +256,21 @@ kinit USERNAME
256256
hive.executeQuery("SELECT * FROM demo").show()
257257
```
258258

259-
:::image type="content" source="./media/apache-hive-warehouse-connector/hive-warehouse-connector-table-before-ranger-policy.png" alt-text="demo table before applying ranger policy." border="true":::
259+
:::image type="content" source="./media/apache-hive-warehouse-connector/hive-warehouse-connector-table-before-ranger-policy.png" alt-text="Screenshot of demo table before applying ranger policy." border="true":::
260260

261261
1. Apply a column masking policy that only shows the last four characters of the column.
262262
1. Go to the Ranger Admin UI at `https://LLAPCLUSTERNAME.azurehdinsight.net/ranger/`.
263263
1. Click on the Hive service for your cluster under **Hive**.
264-
:::image type="content" source="./media/apache-hive-warehouse-connector/hive-warehouse-connector-ranger-service-manager.png" alt-text="ranger service manager." border="true":::
264+
:::image type="content" source="./media/apache-hive-warehouse-connector/hive-warehouse-connector-ranger-service-manager.png" alt-text="Screenshot of ranger service manager." border="true":::
265265
1. Click on the **Masking** tab and then **Add New Policy**
266266

267-
:::image type="content" source="./media/apache-hive-warehouse-connector/hive-warehouse-connector-ranger-hive-policy-list.png" alt-text="hive warehouse connector ranger hive policy list." border="true":::
267+
:::image type="content" source="./media/apache-hive-warehouse-connector/hive-warehouse-connector-ranger-hive-policy-list.png" alt-text="Screenshot of hive warehouse connector ranger hive policy list." border="true":::
268268

269269
1. Provide a desired policy name. Select database: **Default**, Hive table: **demo**, Hive column: **name**, User: **rsadmin2**, Access Types: **select**, and **Partial mask: show last 4** from the **Select Masking Option** menu. Click **Add**.
270-
:::image type="content" source="./media/apache-hive-warehouse-connector/hive-warehouse-connector-ranger-create-policy.png" alt-text="create policy." border="true":::
270+
:::image type="content" source="./media/apache-hive-warehouse-connector/hive-warehouse-connector-ranger-create-policy.png" alt-text="Screenshot of create policy." border="true":::
271271
1. View the table's contents again. After applying the ranger policy, we can see only the last four characters of the column.
272272

273-
:::image type="content" source="./media/apache-hive-warehouse-connector/hive-warehouse-connector-table-after-ranger-policy.png" alt-text="demo table after applying ranger policy." border="true":::
273+
:::image type="content" source="./media/apache-hive-warehouse-connector/hive-warehouse-connector-table-after-ranger-policy.png" alt-text="Screenshot of demo table after applying ranger policy." border="true":::
274274

275275
## Next steps
276276

0 commit comments

Comments
 (0)