File tree Expand file tree Collapse file tree 6 files changed +25
-6
lines changed
job-server-python/src/python/test Expand file tree Collapse file tree 6 files changed +25
-6
lines changed Original file line number Diff line number Diff line change @@ -24,7 +24,7 @@ RUN sbt update
24
24
# add the rest of the code
25
25
COPY . .
26
26
27
- ENV SPARK_HOME /tmp/spark-2.1 .0-bin-hadoop2.7
27
+ ENV SPARK_HOME /tmp/spark-2.2 .0-bin-hadoop2.7
28
28
ENV JAVA_OPTIONS "-Xmx1500m -XX:MaxPermSize=512m -Dakka.test.timefactor=3"
29
29
30
30
CMD ["/usr/src/app/run_tests.sh"]
Original file line number Diff line number Diff line change @@ -119,7 +119,7 @@ Spark Job Server is now included in Datastax Enterprise 4.8!
119
119
| 0.6.1 | 1.5.2 |
120
120
| 0.6.2 | 1.6.1 |
121
121
| 0.7.0 | 1.6.2 |
122
- | 0.8.0-SNAPSHOT | 2.1 .0 |
122
+ | 0.8.0-SNAPSHOT | 2.2 .0 |
123
123
124
124
For release notes, look in the ` notes/ ` directory.
125
125
Original file line number Diff line number Diff line change 1
1
#! /usr/bin/env bash
2
2
set -e
3
- curl -L -o /tmp/spark.tgz http://d3kbcqa49mib13.cloudfront.net/spark-2.1 .0-bin-hadoop2.7.tgz
3
+ curl -L -o /tmp/spark.tgz http://d3kbcqa49mib13.cloudfront.net/spark-2.2 .0-bin-hadoop2.7.tgz
4
4
tar -xvzf /tmp/spark.tgz -C /tmp
Original file line number Diff line number Diff line change
1
+ import errno
2
+ import os
1
3
import unittest
2
4
from pyhocon import ConfigFactory
3
5
from pyspark import SparkConf , SparkContext
6
8
from py4j .java_gateway import java_import
7
9
8
10
11
+ def silentremove (filename ):
12
+ try :
13
+ os .remove (filename )
14
+ except OSError as exc :
15
+ if exc .errno != errno .ENOENT :
16
+ raise
17
+
18
+
9
19
class WordCountSparkJob (SparkJob ):
10
20
"""
11
21
Simple example of a SparkContext job for use in tests
@@ -56,6 +66,11 @@ class TestSJSApi(unittest.TestCase):
56
66
57
67
def setUp (self ):
58
68
conf = SparkConf ().setAppName ('test' ).setMaster ('local[*]' )
69
+ pwd = os .path .dirname (os .path .realpath (__file__ ))
70
+ metastore_dir = os .path .abspath (os .path .join (pwd , '..' ,
71
+ 'metastore_db' ))
72
+ silentremove (os .path .join (metastore_dir , "dbex.lck" ))
73
+ silentremove (os .path .join (metastore_dir , "db.lck" ))
59
74
self .sc = SparkContext (conf = conf )
60
75
self .jvm = self .sc ._gateway .jvm
61
76
java_import (self .jvm , "org.apache.spark.sql.*" )
Original file line number Diff line number Diff line change @@ -34,6 +34,8 @@ object Dependencies {
34
34
)
35
35
36
36
lazy val sparkExtraDeps = Seq (
37
+ " org.apache.derby" % " derby" % derby % Provided excludeAll(excludeNettyIo, excludeQQ),
38
+ " org.apache.hadoop" % " hadoop-client" % hadoop % Provided excludeAll(excludeNettyIo, excludeQQ),
37
39
" org.apache.spark" %% " spark-mllib" % spark % Provided excludeAll(excludeNettyIo, excludeQQ),
38
40
" org.apache.spark" %% " spark-sql" % spark % Provided excludeAll(excludeNettyIo, excludeQQ),
39
41
" org.apache.spark" %% " spark-streaming" % spark % Provided excludeAll(excludeNettyIo, excludeQQ),
Original file line number Diff line number Diff line change 1
1
import scala .util .Properties .isJavaAtLeast
2
2
3
3
object Versions {
4
- lazy val spark = sys.env.getOrElse(" SPARK_VERSION" , " 2.1 .0" )
4
+ lazy val spark = sys.env.getOrElse(" SPARK_VERSION" , " 2.2 .0" )
5
5
6
6
lazy val akka = " 2.4.9"
7
- lazy val cassandra = " 3.0.3 "
7
+ lazy val cassandra = " 3.3.0 "
8
8
lazy val cassandraUnit = " 2.2.2.1"
9
9
lazy val commons = " 1.4"
10
+ lazy val derby = " 10.12.1.1"
10
11
lazy val flyway = " 3.2.1"
12
+ lazy val hadoop = " 2.7.3"
11
13
lazy val h2 = " 1.3.176"
12
14
lazy val java = sys.env.getOrElse(" JAVA_VERSION" , " 8-jdk" )
13
15
lazy val jodaConvert = " 1.8.1"
14
16
lazy val jodaTime = " 2.9.3"
15
17
lazy val logback = " 1.0.7"
16
18
lazy val mesos = sys.env.getOrElse(" MESOS_VERSION" , " 1.0.0-2.0.89.ubuntu1404" )
17
19
lazy val metrics = " 2.2.0"
18
- lazy val netty = " 4.0.42 .Final"
20
+ lazy val netty = " 4.0.44 .Final"
19
21
lazy val postgres = " 9.4.1209"
20
22
lazy val mysql = " 5.1.42"
21
23
lazy val py4j = " 0.10.4"
You can’t perform that action at this time.
0 commit comments