Skip to content

Commit 36d6628

Browse files
authored
Merge pull request #323 from ldbc/file-uri-fixes
File uri fixes
2 parents dc5211e + 7587594 commit 36d6628

File tree

14 files changed

+63
-186
lines changed

14 files changed

+63
-186
lines changed

pom.xml

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -72,11 +72,6 @@
7272
<artifactId>shapeless_${scala.compat.version}</artifactId>
7373
<version>2.3.3</version>
7474
</dependency>
75-
<dependency>
76-
<groupId>com.github.pathikrit</groupId>
77-
<artifactId>better-files_${scala.compat.version}</artifactId>
78-
<version>3.8.0</version>
79-
</dependency>
8075
<dependency>
8176
<groupId>com.github.scopt</groupId>
8277
<artifactId>scopt_${scala.compat.version}</artifactId>

src/main/java/ldbc/snb/datagen/hadoop/serializer/HadoopStaticSerializer.java

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -47,6 +47,7 @@ Linked Data Benchmark Council (http://www.ldbcouncil.org)
4747
import org.apache.hadoop.conf.Configuration;
4848
import org.apache.hadoop.fs.FileSystem;
4949

50+
import java.net.URI;
5051
import java.util.Iterator;
5152
import java.util.Set;
5253
import java.util.TreeSet;
@@ -71,7 +72,7 @@ public HadoopStaticSerializer(GeneratorConfiguration conf, Configuration hadoopC
7172

7273
public void run() {
7374
try {
74-
FileSystem fs = FileSystem.get(hadoopConf);
75+
FileSystem fs = FileSystem.get(new URI(conf.getOutputDir()), hadoopConf);
7576
staticSerializer = new StaticSerializer[numPartitions];
7677
for (int i = 0; i < numPartitions; ++i) {
7778
staticSerializer[i] = new StaticSerializer();

src/main/java/ldbc/snb/datagen/hadoop/sorting/HadoopCreationTimeSorter.java

Lines changed: 0 additions & 61 deletions
This file was deleted.

src/main/java/ldbc/snb/datagen/hadoop/sorting/HadoopDeletionTimeSorter.java

Lines changed: 0 additions & 67 deletions
This file was deleted.

src/main/java/ldbc/snb/datagen/hadoop/sorting/SortPartitioner.java

Lines changed: 0 additions & 26 deletions
This file was deleted.

src/main/java/ldbc/snb/datagen/hadoop/sorting/SortReducer.java

Lines changed: 0 additions & 16 deletions
This file was deleted.

src/main/scala/ldbc/snb/datagen/factors/io/package.scala

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ package ldbc.snb.datagen.factors
33
import ldbc.snb.datagen.io.{PathComponent, Writer}
44
import ldbc.snb.datagen.util.Logging
55
import ldbc.snb.datagen.model.{GraphLike, Mode}
6-
import better.files._
6+
import ldbc.snb.datagen.syntax._
77
import ldbc.snb.datagen.io.dataframes.DataFrameSink
88

99
package object io {
@@ -16,7 +16,7 @@ package object io {
1616
override type CoRet = FactorTable[M]
1717

1818
override def write(self: FactorTable[M], sink: FactorTableSink): Unit = {
19-
val p = (sink.path / "factors" / sink.format / PathComponent[GraphLike[M]].path(self.source) / self.name).toString()
19+
val p = (sink.path / "factors" / sink.format / PathComponent[GraphLike[M]].path(self.source) / self.name).toString
2020
self.data.coalesce(1).write(DataFrameSink(p, sink.format))
2121
log.info(s"Factor table ${self.name} written")
2222
}

src/main/scala/ldbc/snb/datagen/generation/serializer/SparkActivitySerializer.scala

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,6 +12,7 @@ import org.apache.spark.TaskContext
1212
import org.apache.spark.rdd.RDD
1313
import org.apache.spark.sql.SparkSession
1414

15+
import java.net.URI
1516
import java.util
1617
import java.util.Collections
1718
import java.util.function.Consumer
@@ -35,7 +36,7 @@ object SparkActivitySerializer {
3536
val hadoopConf = serializableHadoopConf.value
3637
val buildDir = conf.getOutputDir
3738

38-
val fs = FileSystem.get(hadoopConf)
39+
val fs = FileSystem.get(new URI(buildDir), hadoopConf)
3940
fs.mkdirs(new Path(buildDir))
4041

4142
val dynamicActivitySerializer = new DynamicActivitySerializer()

src/main/scala/ldbc/snb/datagen/generation/serializer/SparkPersonSerializer.scala

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,6 +10,8 @@ import org.apache.spark.TaskContext
1010
import org.apache.spark.rdd.RDD
1111
import org.apache.spark.sql.SparkSession
1212

13+
import java.net.URI
14+
1315
object SparkPersonSerializer {
1416

1517
def apply(
@@ -28,7 +30,7 @@ object SparkPersonSerializer {
2830
val partitionId = TaskContext.getPartitionId()
2931
val buildDir = conf.getOutputDir
3032

31-
val fs = FileSystem.get(hadoopConf)
33+
val fs = FileSystem.get(new URI(buildDir), hadoopConf)
3234
fs.mkdirs(new Path(buildDir))
3335

3436
dynamicPersonSerializer.initialize(

src/main/scala/ldbc/snb/datagen/io/dataframes.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
package ldbc.snb.datagen.io
22

3-
import ldbc.snb.datagen.syntax.fluentSyntaxOps
3+
import ldbc.snb.datagen.syntax._
44
import org.apache.spark.sql.types.StructType
55
import org.apache.spark.sql.{DataFrame, SaveMode, SparkSession}
66

0 commit comments

Comments
 (0)