Skip to content

Commit 101c8f7

Browse files
authored
Update apache-hive-warehouse-connector.md
1 parent a5bc88f commit 101c8f7

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

articles/hdinsight/interactive-query/apache-hive-warehouse-connector.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ The Hive Warehouse Connector allows you to take advantage of the unique features
1616

1717
Apache Hive offers support for database transactions that are Atomic, Consistent, Isolated, and Durable (ACID). For more information on ACID and transactions in Hive, see [Hive Transactions](https://cwiki.apache.org/confluence/display/Hive/Hive+Transactions). Hive also offers detailed security controls through Apache Ranger and Low Latency Analytical Processing (LLAP) not available in Apache Spark.
1818

19-
Apache Spark, has a Structured Streaming API that gives streaming capabilities not available in Apache Hive. Beginning with HDInsight 4.0, Apache Spark 2.3.1 & above, and Apache Hive 3.1.0 have separate metastore catalogs, which make interoperability difficult.
19+
Apache Spark has a Structured Streaming API that gives streaming capabilities not available in Apache Hive. Beginning with HDInsight 4.0, Apache Spark 2.3.1 & above, and Apache Hive 3.1.0 have separate metastore catalogs, which make interoperability difficult.
2020

2121
The Hive Warehouse Connector (HWC) makes it easier to use Spark and Hive together. The HWC library loads data from LLAP daemons to Spark executors in parallel. This process makes it more efficient and adaptable than a standard JDBC connection from Spark to Hive. This brings out two different execution modes for HWC:
2222
> - Hive JDBC mode via HiveServer2
@@ -199,7 +199,7 @@ Once you build the scala/java code along with the dependencies into an assembly
199199
/<APP_JAR_PATH>/myHwcAppProject.jar
200200
```
201201
202-
This utility is also used when we have written the entire application in pySpark and packaged into py files (Python), so that we can submit the entire code to Spark cluster for execution.
202+
This utility is also used when we have written the entire application in pySpark and packaged into `.py` files (Python), so that we can submit the entire code to Spark cluster for execution.
203203
204204
For Python applications, pass a .py file in the place of `/<APP_JAR_PATH>/myHwcAppProject.jar`, and add the below configuration (Python .zip) file to the search path with `--py-files`.
205205

0 commit comments

Comments
 (0)