Skip to content

Commit 07c75c4

Browse files
Merge pull request #262412 from v-akarnase/patch-52
Update apache-hive-warehouse-connector.md
2 parents 6fbf252 + 101c8f7 commit 07c75c4

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

articles/hdinsight/interactive-query/apache-hive-warehouse-connector.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ author: reachnijel
55
ms.author: nijelsf
66
ms.service: hdinsight
77
ms.topic: how-to
8-
ms.date: 12/09/2022
8+
ms.date: 01/04/2024
99
---
1010

1111
# Integrate Apache Spark and Apache Hive with Hive Warehouse Connector in Azure HDInsight
@@ -16,7 +16,7 @@ The Hive Warehouse Connector allows you to take advantage of the unique features
1616

1717
Apache Hive offers support for database transactions that are Atomic, Consistent, Isolated, and Durable (ACID). For more information on ACID and transactions in Hive, see [Hive Transactions](https://cwiki.apache.org/confluence/display/Hive/Hive+Transactions). Hive also offers detailed security controls through Apache Ranger and Low Latency Analytical Processing (LLAP) not available in Apache Spark.
1818

19-
Apache Spark, has a Structured Streaming API that gives streaming capabilities not available in Apache Hive. Beginning with HDInsight 4.0, Apache Spark 2.3.1 & above, and Apache Hive 3.1.0 have separate metastore catalogs, which make interoperability difficult.
19+
Apache Spark has a Structured Streaming API that gives streaming capabilities not available in Apache Hive. Beginning with HDInsight 4.0, Apache Spark 2.3.1 & above, and Apache Hive 3.1.0 have separate metastore catalogs, which make interoperability difficult.
2020

2121
The Hive Warehouse Connector (HWC) makes it easier to use Spark and Hive together. The HWC library loads data from LLAP daemons to Spark executors in parallel. This process makes it more efficient and adaptable than a standard JDBC connection from Spark to Hive. This brings out two different execution modes for HWC:
2222
> - Hive JDBC mode via HiveServer2
@@ -199,7 +199,7 @@ Once you build the scala/java code along with the dependencies into an assembly
199199
/<APP_JAR_PATH>/myHwcAppProject.jar
200200
```
201201
202-
This utility is also used when we have written the entire application in pySpark and packaged into py files (Python), so that we can submit the entire code to Spark cluster for execution.
202+
This utility is also used when we have written the entire application in pySpark and packaged into `.py` files (Python), so that we can submit the entire code to Spark cluster for execution.
203203
204204
For Python applications, pass a .py file in the place of `/<APP_JAR_PATH>/myHwcAppProject.jar`, and add the below configuration (Python .zip) file to the search path with `--py-files`.
205205

0 commit comments

Comments
 (0)