Skip to content

Commit 4e7f4f0

Browse files
dansanduleacrobert3005
authored andcommitted
Properly remove hive from modules (apache-spark-on-k8s#338)
1 parent 5774deb commit 4e7f4f0

File tree

10 files changed

+27
-57
lines changed

10 files changed

+27
-57
lines changed

.circleci/config.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -85,7 +85,7 @@ jobs:
8585
- build-binaries-{{ checksum "build/mvn" }}-{{ checksum "build/sbt" }}
8686
- build-binaries-
8787
- run: |
88-
./build/mvn -T1C -DskipTests -Phadoop-cloud -Phadoop-palantir -Pkinesis-asl -Pkubernetes -Pyarn -Phive -Psparkr install \
88+
./build/mvn -T1C -DskipTests -Phadoop-cloud -Phadoop-palantir -Pkinesis-asl -Pkubernetes -Pyarn -Psparkr install \
8989
| tee -a "/tmp/mvn-install.log"
9090
- store_artifacts:
9191
path: /tmp/mvn-install.log

dev/deps/spark-deps-hadoop-palantir

Lines changed: 0 additions & 25 deletions
Original file line numberDiff line numberDiff line change
@@ -1,14 +1,9 @@
1-
JavaEWAH-0.3.2.jar
21
RoaringBitmap-0.6.43.jar
3-
ST4-4.0.4.jar
42
activation-1.1.1.jar
53
aircompressor-0.8.jar
6-
antlr-2.7.7.jar
7-
antlr-runtime-3.4.jar
84
antlr4-runtime-4.7.jar
95
aopalliance-1.0.jar
106
aopalliance-repackaged-2.5.0-b32.jar
11-
apache-log4j-extras-1.2.17.jar
127
apacheds-i18n-2.0.0-M15.jar
138
apacheds-kerberos-codec-2.0.0-M15.jar
149
api-asn1-api-1.0.0-M20.jar
@@ -27,12 +22,8 @@ aws-java-sdk-s3-1.11.45.jar
2722
base64-2.3.8.jar
2823
bcpkix-jdk15on-1.58.jar
2924
bcprov-jdk15on-1.58.jar
30-
bonecp-0.8.0.RELEASE.jar
3125
breeze-macros_2.11-0.13.2.jar
3226
breeze_2.11-0.13.2.jar
33-
calcite-avatica-1.2.0-incubating.jar
34-
calcite-core-1.2.0-incubating.jar
35-
calcite-linq4j-1.2.0-incubating.jar
3627
chill-java-0.8.4.jar
3728
chill_2.11-0.8.4.jar
3829
classmate-1.1.0.jar
@@ -45,7 +36,6 @@ commons-compiler-3.0.8.jar
4536
commons-compress-1.8.1.jar
4637
commons-configuration-1.6.jar
4738
commons-crypto-1.0.0.jar
48-
commons-dbcp-1.4.jar
4939
commons-digester-1.8.jar
5040
commons-httpclient-3.1.jar
5141
commons-io-2.4.jar
@@ -62,17 +52,12 @@ core-1.1.2.jar
6252
curator-client-2.7.1.jar
6353
curator-framework-2.7.1.jar
6454
curator-recipes-2.7.1.jar
65-
datanucleus-api-jdo-3.2.6.jar
66-
datanucleus-core-3.2.10.jar
67-
datanucleus-rdbms-3.2.9.jar
68-
derby-10.12.1.1.jar
6955
dropwizard-jackson-0.9.2.jar
7056
dropwizard-lifecycle-0.9.2.jar
7157
dropwizard-metrics-0.9.2.jar
7258
dropwizard-metrics-influxdb-1.1.8.jar
7359
dropwizard-util-0.9.2.jar
7460
dropwizard-validation-0.9.2.jar
75-
eigenbase-properties-1.1.5.jar
7661
fastutil-7.0.13.jar
7762
flatbuffers-1.2.0-3f79e055.jar
7863
generex-1.0.1.jar
@@ -134,12 +119,10 @@ javax.inject-1.jar
134119
javax.inject-2.5.0-b32.jar
135120
javax.servlet-api-3.1.0.jar
136121
javax.ws.rs-api-2.0.1.jar
137-
javolution-5.5.1.jar
138122
jaxb-api-2.2.2.jar
139123
jboss-logging-3.2.1.Final.jar
140124
jcip-annotations-1.0.jar
141125
jcl-over-slf4j-1.7.25.jar
142-
jdo-api-3.0.1.jar
143126
jersey-client-2.25.1.jar
144127
jersey-common-2.25.1.jar
145128
jersey-container-servlet-2.25.1.jar
@@ -154,23 +137,19 @@ jetty-sslengine-6.1.26.jar
154137
jetty-util-6.1.26.jar
155138
jmespath-java-1.0.jar
156139
joda-time-2.9.9.jar
157-
jodd-core-3.5.2.jar
158140
json-smart-1.1.1.jar
159141
json4s-ast_2.11-3.5.3.jar
160142
json4s-core_2.11-3.5.3.jar
161143
json4s-jackson_2.11-3.5.3.jar
162144
json4s-scalap_2.11-3.5.3.jar
163145
jsp-api-2.1.jar
164146
jsr305-3.0.1.jar
165-
jta-1.1.jar
166147
jtransforms-2.4.0.jar
167148
jul-to-slf4j-1.7.25.jar
168149
kryo-shaded-3.0.3.jar
169150
kubernetes-client-3.0.0.jar
170151
kubernetes-model-2.0.0.jar
171152
leveldbjni-all-1.8.jar
172-
libfb303-0.9.3.jar
173-
libthrift-0.9.3.jar
174153
log4j-1.2.17.jar
175154
logging-interceptor-3.9.1.jar
176155
lz4-java-1.4.0.jar
@@ -201,7 +180,6 @@ parquet-common-1.9.1-palantir3.jar
201180
parquet-encoding-1.9.1-palantir3.jar
202181
parquet-format-2.3.1.jar
203182
parquet-hadoop-1.9.1-palantir3.jar
204-
parquet-hadoop-bundle-1.6.0.jar
205183
parquet-jackson-1.9.1-palantir3.jar
206184
protobuf-java-2.5.0.jar
207185
py4j-0.10.6.jar
@@ -216,14 +194,11 @@ shapeless_2.11-2.3.2.jar
216194
slf4j-api-1.7.25.jar
217195
slf4j-log4j12-1.7.25.jar
218196
snakeyaml-1.15.jar
219-
snappy-0.2.jar
220197
snappy-java-1.1.7.1.jar
221198
spire-macros_2.11-0.13.0.jar
222199
spire_2.11-0.13.0.jar
223200
stax-api-1.0-2.jar
224-
stax-api-1.0.1.jar
225201
stream-2.7.0.jar
226-
stringtemplate-3.2.1.jar
227202
univocity-parsers-2.5.9.jar
228203
validation-api-1.1.0.Final.jar
229204
xbean-asm5-shaded-4.4.jar

dev/publish-local.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
set -euo pipefail
44
version=$(git describe --tags --first-parent)
55

6-
PALANTIR_FLAGS=(-Phadoop-cloud -Phadoop-palantir -Pkinesis-asl -Pkubernetes -Phive -Pyarn -Psparkr)
6+
PALANTIR_FLAGS=(-Phadoop-cloud -Phadoop-palantir -Pkinesis-asl -Pkubernetes -Pyarn -Psparkr)
77

88
MVN_LOCAL="~/.m2/repository"
99

dev/publish_functions.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22

33
set -euo pipefail
44

5-
PALANTIR_FLAGS=(-Phadoop-cloud -Phadoop-palantir -Pkinesis-asl -Pkubernetes -Phive -Pyarn -Psparkr)
5+
PALANTIR_FLAGS=(-Phadoop-cloud -Phadoop-palantir -Pkinesis-asl -Pkubernetes -Pyarn -Psparkr)
66

77
get_version() {
88
git describe --tags --first-parent

dev/sparktestsupport/modules.py

Lines changed: 21 additions & 21 deletions
Original file line numberDiff line numberDiff line change
@@ -124,28 +124,28 @@ def __hash__(self):
124124
)
125125

126126

127-
hive = Module(
128-
name="hive",
129-
dependencies=[sql],
130-
source_file_regexes=[
131-
"sql/hive/",
132-
"bin/spark-sql",
133-
],
134-
build_profile_flags=[
135-
"-Phive",
136-
],
137-
sbt_test_goals=[
138-
"hive/test",
139-
],
140-
test_tags=[
141-
"org.apache.spark.tags.ExtendedHiveTest"
142-
]
143-
)
127+
# hive = Module(
128+
# name="hive",
129+
# dependencies=[sql],
130+
# source_file_regexes=[
131+
# "sql/hive/",
132+
# "bin/spark-sql",
133+
# ],
134+
# build_profile_flags=[
135+
# "-Phive",
136+
# ],
137+
# sbt_test_goals=[
138+
# "hive/test",
139+
# ],
140+
# test_tags=[
141+
# "org.apache.spark.tags.ExtendedHiveTest"
142+
# ]
143+
# )
144144

145145

146146
repl = Module(
147147
name="repl",
148-
dependencies=[hive],
148+
dependencies=[],
149149
source_file_regexes=[
150150
"repl/",
151151
],
@@ -351,7 +351,7 @@ def __hash__(self):
351351

352352
examples = Module(
353353
name="examples",
354-
dependencies=[graphx, mllib, streaming, hive],
354+
dependencies=[graphx, mllib, streaming],
355355
source_file_regexes=[
356356
"examples/",
357357
],
@@ -384,7 +384,7 @@ def __hash__(self):
384384

385385
pyspark_sql = Module(
386386
name="pyspark-sql",
387-
dependencies=[pyspark_core, hive],
387+
dependencies=[pyspark_core],
388388
source_file_regexes=[
389389
"python/pyspark/sql"
390390
],
@@ -486,7 +486,7 @@ def __hash__(self):
486486

487487
sparkr = Module(
488488
name="sparkr",
489-
dependencies=[hive, mllib],
489+
dependencies=[mllib],
490490
source_file_regexes=[
491491
"R/",
492492
],

dev/test-dependencies.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@ export LC_ALL=C
2929
# TODO: This would be much nicer to do in SBT, once SBT supports Maven-style resolution.
3030

3131
# NOTE: These should match those in the release publishing script
32-
HADOOP2_MODULE_PROFILES="-Phadoop-cloud -Pkubernetes -Pyarn -Pflume -Phive"
32+
HADOOP2_MODULE_PROFILES="-Phadoop-cloud -Pkubernetes -Pkinesis-asl -Pyarn"
3333
MVN="build/mvn"
3434
HADOOP_PROFILES=(
3535
hadoop-palantir

dists/hadoop-palantir/pom.xml

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -195,10 +195,6 @@
195195
<artifactId>guava</artifactId>
196196
<scope>${hadoop.deps.scope}</scope>
197197
</dependency>
198-
<dependency>
199-
<groupId>org.apache.spark</groupId>
200-
<artifactId>spark-hive_${scala.binary.version}</artifactId>
201-
</dependency>
202198
<dependency>
203199
<groupId>org.apache.spark</groupId>
204200
<artifactId>spark-yarn_${scala.binary.version}</artifactId>

project/SparkBuild.scala

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -559,8 +559,6 @@ object SQL {
559559
object Hive {
560560

561561
lazy val settings = Seq(
562-
// Skip all hive tests because we don't use hive.
563-
definedTests in Test := Nil,
564562
// Specially disable assertions since some Hive tests fail them
565563
javaOptions in Test := (javaOptions in Test).value.filterNot(_ == "-ea"),
566564
// Supporting all SerDes requires us to depend on deprecated APIs, so we turn off the warnings

python/pyspark/sql/readwriter.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -954,7 +954,7 @@ def _test():
954954
globs = pyspark.sql.readwriter.__dict__.copy()
955955
sc = SparkContext('local[4]', 'PythonTest')
956956
try:
957-
spark = SparkSession.builder.enableHiveSupport().getOrCreate()
957+
spark = SparkSession.builder.getOrCreate()
958958
except py4j.protocol.Py4JError:
959959
spark = SparkSession(sc)
960960

python/pyspark/sql/tests.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -2923,6 +2923,7 @@ def test_create_dateframe_from_pandas_with_dst(self):
29232923

29242924
class HiveSparkSubmitTests(SparkSubmitTests):
29252925

2926+
@unittest.skip("This distrbituion doesn't ship hive")
29262927
def test_hivecontext(self):
29272928
# This test checks that HiveContext is using Hive metastore (SPARK-16224).
29282929
# It sets a metastore url and checks if there is a derby dir created by

0 commit comments

Comments
 (0)