Skip to content

Commit 92bad96

Browse files
Merge pull request #1 from techmonad/version_update
Update spark and elasticsearch version
2 parents d935b9b + 74adeec commit 92bad96

File tree

2 files changed

+11
-10
lines changed

2 files changed

+11
-10
lines changed

build.sbt

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -2,13 +2,13 @@ name := "spark-data-pipeline"
22

33
version := "1.0"
44

5-
scalaVersion := "2.11.11"
5+
scalaVersion := "2.11.12"
66

77

88
libraryDependencies ++= Seq(
9-
"org.apache.spark" %% "spark-core" % "2.3.1",
9+
"org.apache.spark" %% "spark-core" % "2.4.0",
1010
"com.univocity" % "univocity-parsers" % "2.6.4",
11-
"org.elasticsearch" %% "elasticsearch-spark-20" % "6.3.0",
11+
"org.elasticsearch" %% "elasticsearch-spark-20" % "6.6.0",
1212
"edu.stanford.nlp" % "stanford-corenlp" % "3.6.0" artifacts(Artifact("stanford-corenlp", "models"), Artifact("stanford-corenlp")),
1313
"ch.qos.logback" % "logback-classic" % "1.2.3",
1414
"org.json4s" %% "json4s-native" % "3.5.4",

src/main/scala/com/techmonad/pipeline/reader/CSVReader.scala

Lines changed: 8 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -10,13 +10,14 @@ object CSVReader extends TryHelper {
1010

1111
def read(url: String, delimiter: Char = ',')(implicit sc: SparkContext): RDD[Record] = {
1212
val headers = sc.textFile(url).first()
13-
sc.textFile(url).mapPartitionsWithIndex {
14-
case (index, itr) =>
15-
if (index == 0)
16-
readFile(itr.drop(1), headers)
17-
else
18-
readFile(itr, headers)
19-
}
13+
sc.textFile(url)
14+
.mapPartitionsWithIndex {
15+
case (index, itr) =>
16+
if (index == 0)
17+
readFile(itr.drop(1), headers)
18+
else
19+
readFile(itr, headers)
20+
}
2021
}
2122

2223

0 commit comments

Comments
 (0)