Skip to content

Commit c46d56d

Browse files
author
Robert Kruszewski
committed
Resolve conflicts
1 parent a8b09ac commit c46d56d

File tree

67 files changed

+244
-2957
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

67 files changed

+244
-2957
lines changed

FORK.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,6 @@
66
* [SPARK-17059](https://issues.apache.org/jira/browse/SPARK-17059) - Allow FileFormat to specify partition pruning strategy via splits
77
* [SPARK-24345](https://issues.apache.org/jira/browse/SPARK-24345) - Improve ParseError stop location when offending symbol is a token
88
* [SPARK-23795](https://issues.apache.org/jira/browse/SPARK-23795) - Make AbstractLauncher#self() protected
9-
* [SPARK-23153](https://issues.apache.org/jira/browse/SPARK-23153) - Support application dependencies in submission client's local file system
109
* [SPARK-18079](https://issues.apache.org/jira/browse/SPARK-18079) - CollectLimitExec.executeToIterator should perform per-partition limits
1110

1211
* [SPARK-15777](https://issues.apache.org/jira/browse/SPARK-15777) (Partial fix) - Catalog federation

core/pom.xml

Lines changed: 5 additions & 36 deletions
Original file line numberDiff line numberDiff line change
@@ -387,48 +387,17 @@
387387
<type>test-jar</type>
388388
<scope>test</scope>
389389
</dependency>
390-
391-
<dependency>
392-
<groupId>org.apache.commons</groupId>
393-
<artifactId>commons-crypto</artifactId>
394-
</dependency>
395-
396-
<<<<<<< HEAD
397-
<!--
398-
The following dependencies are depended upon in HiveCredentialProvider, but are only executed if Hive is enabled in
399-
the user's Hadoop configuration. So in order to prevent spark-core from depending on Hive, these deps have been
400-
placed in the "provided" scope, rather than the "compile" scope, and NoClassDefFoundError exceptions are handled
401-
when the user has not explicitly compiled with the Hive module.
402-
-->
403-
<dependency>
404-
<groupId>${hive.group}</groupId>
405-
<artifactId>hive-exec</artifactId>
406-
<scope>provided</scope>
407-
</dependency>
408-
<dependency>
409-
<groupId>${hive.group}</groupId>
410-
<artifactId>hive-metastore</artifactId>
411-
<scope>provided</scope>
412-
</dependency>
413-
<dependency>
414-
<groupId>org.apache.thrift</groupId>
415-
<artifactId>libthrift</artifactId>
416-
<scope>provided</scope>
417-
</dependency>
418-
<dependency>
419-
<groupId>org.apache.thrift</groupId>
420-
<artifactId>libfb303</artifactId>
421-
<scope>provided</scope>
422-
</dependency>
423-
424390
<dependency>
425391
<groupId>com.palantir.safe-logging</groupId>
426392
<artifactId>safe-logging</artifactId>
427393
<version>1.5.1</version>
428394
</dependency>
429395

430-
=======
431-
>>>>>>> master
396+
<dependency>
397+
<groupId>org.apache.commons</groupId>
398+
<artifactId>commons-crypto</artifactId>
399+
</dependency>
400+
432401
</dependencies>
433402
<build>
434403
<outputDirectory>target/scala-${scala.binary.version}/classes</outputDirectory>

core/src/main/scala/org/apache/spark/SparkContext.scala

Lines changed: 0 additions & 21 deletions
Original file line numberDiff line numberDiff line change
@@ -440,11 +440,7 @@ class SparkContext(config: SparkConf) extends SafeLogging {
440440
_statusTracker = new SparkStatusTracker(this, _statusStore)
441441

442442
_progressBar =
443-
<<<<<<< HEAD
444443
if (_conf.get(UI_SHOW_CONSOLE_PROGRESS) && !safeLogIsInfoEnabled) {
445-
=======
446-
if (_conf.get(UI_SHOW_CONSOLE_PROGRESS)) {
447-
>>>>>>> master
448444
Some(new ConsoleProgressBar(this))
449445
} else {
450446
None
@@ -2488,35 +2484,18 @@ object SparkContext extends SafeLogging {
24882484
Option(activeContext.get()).filter(_ ne sc).foreach { ctx =>
24892485
val errMsg = "Only one SparkContext should be running in this JVM (see SPARK-2243)." +
24902486
s"The currently running SparkContext was created at:\n${ctx.creationSite.longForm}"
2491-
<<<<<<< HEAD
2492-
val exception = new SparkException(errMsg)
2493-
if (allowMultipleContexts) {
2494-
safeLogWarning("Multiple running SparkContexts detected in the same JVM!", exception)
2495-
} else {
2496-
throw exception
2497-
}
2498-
=======
24992487
throw new SparkException(errMsg)
2500-
>>>>>>> master
25012488
}
25022489

25032490
contextBeingConstructed.filter(_ ne sc).foreach { otherContext =>
25042491
// Since otherContext might point to a partially-constructed context, guard against
25052492
// its creationSite field being null:
25062493
val otherContextCreationSite =
25072494
Option(otherContext.creationSite).map(_.longForm).getOrElse("unknown location")
2508-
<<<<<<< HEAD
25092495
safeLogWarning("Another SparkContext is being constructed (or threw an exception in its" +
25102496
" constructor). This may indicate an error, since only one SparkContext may be" +
25112497
" running in this JVM (see SPARK-2243).",
25122498
UnsafeArg.of("otherContextCreationSite", otherContextCreationSite))
2513-
=======
2514-
val warnMsg = "Another SparkContext is being constructed (or threw an exception in its" +
2515-
" constructor). This may indicate an error, since only one SparkContext should be" +
2516-
" running in this JVM (see SPARK-2243)." +
2517-
s" The other SparkContext was created at:\n$otherContextCreationSite"
2518-
logWarning(warnMsg)
2519-
>>>>>>> master
25202499
}
25212500
}
25222501
}

core/src/main/scala/org/apache/spark/api/r/RRunner.scala

Lines changed: 7 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -349,21 +349,17 @@ private[r] object RRunner {
349349
// "spark.sparkr.r.command" is deprecated and replaced by "spark.r.command",
350350
// but kept here for backward compatibility.
351351
val sparkConf = SparkEnv.get.conf
352-
<<<<<<< HEAD
353-
val requestedRCommand = Provenance.fromConf("spark.r.command")
354-
.orElse(Provenance.fromConf("spark.sparkr.r.command"))
352+
val requestedRCommand = Provenance.fromConfOpt(sparkConf, R_COMMAND)
353+
.getOrElse(Provenance.fromConf(sparkConf, SPARKR_COMMAND))
355354
val condaEnv = condaSetupInstructions.map(CondaEnvironmentManager.getOrCreateCondaEnvironment)
356355
val rCommand = condaEnv.map { conda =>
357-
requestedRCommand.foreach(exec => sys.error(s"It's forbidden to set the r executable " +
358-
s"when using conda, but found: $exec"))
356+
if (requestedRCommand.value != SPARKR_COMMAND.defaultValue.get) {
357+
sys.error(s"It's forbidden to set the r executable " +
358+
s"when using conda, but found: ${requestedRCommand.value}")
359+
}
359360

360361
conda.condaEnvDir + "/bin/Rscript"
361-
}.orElse(requestedRCommand.map(_.value))
362-
.getOrElse("Rscript")
363-
=======
364-
var rCommand = sparkConf.get(SPARKR_COMMAND)
365-
rCommand = sparkConf.get(R_COMMAND).orElse(Some(rCommand)).get
366-
>>>>>>> master
362+
}.getOrElse(requestedRCommand.value)
367363

368364
val rConnectionTimeout = sparkConf.get(R_BACKEND_CONNECTION_TIMEOUT)
369365
val rOptions = "--vanilla"

core/src/main/scala/org/apache/spark/broadcast/TorrentBroadcast.scala

Lines changed: 5 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -29,11 +29,8 @@ import scala.util.Random
2929
import com.palantir.logsafe.SafeArg
3030

3131
import org.apache.spark._
32-
<<<<<<< HEAD
3332
import org.apache.spark.internal.SafeLogging
34-
=======
35-
import org.apache.spark.internal.{config, Logging}
36-
>>>>>>> master
33+
import org.apache.spark.internal.config
3734
import org.apache.spark.io.CompressionCodec
3835
import org.apache.spark.serializer.Serializer
3936
import org.apache.spark.storage._
@@ -244,14 +241,11 @@ private[spark] class TorrentBroadcast[T: ClassTag](obj: T, id: Long)
244241
throw new SparkException(s"Failed to get locally stored broadcast data: $broadcastId")
245242
}
246243
case None =>
247-
<<<<<<< HEAD
248-
safeLogInfo("Started reading broadcast variable",
249-
SafeArg.of("id", id))
250-
=======
251244
val estimatedTotalSize = Utils.bytesToString(numBlocks * blockSize)
252-
logInfo(s"Started reading broadcast variable $id with $numBlocks pieces " +
253-
s"(estimated total size $estimatedTotalSize)")
254-
>>>>>>> master
245+
safeLogInfo("Started reading broadcast variable",
246+
SafeArg.of("id", id),
247+
SafeArg.of("numBlocks", numBlocks),
248+
SafeArg.of("estimatedTotalSize", estimatedTotalSize))
255249
val startTimeMs = System.currentTimeMillis()
256250
val blocks = readBlocks()
257251
safeLogInfo("Reading broadcast variable finished",

core/src/main/scala/org/apache/spark/deploy/Common.scala

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,10 +30,14 @@ object Common {
3030
sys.props.get(sysPropName).map(Provenance(s"Spark config $sysPropName", _))
3131
}
3232

33-
def fromConf(sparkConf: SparkConf, conf: ConfigEntry[Option[String]]): Option[Provenance] = {
33+
def fromConfOpt(sparkConf: SparkConf, conf: ConfigEntry[Option[String]]): Option[Provenance] = {
3434
sparkConf.get(conf).map(Provenance(s"Spark config ${conf.key}", _))
3535
}
3636

37+
def fromConf(sparkConf: SparkConf, conf: ConfigEntry[String]): Provenance = {
38+
Provenance(s"Spark config ${conf.key}", sparkConf.get(conf))
39+
}
40+
3741
def fromConf(sparkConf: SparkConf, key: String): Option[Provenance] = {
3842
sparkConf.getOption(key).map(Provenance(s"Spark config $key", _))
3943
}

core/src/main/scala/org/apache/spark/deploy/PythonRunner.scala

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -46,8 +46,8 @@ object PythonRunner extends CondaRunner with Logging {
4646
val otherArgs = args.slice(2, args.length)
4747
val sparkConf = new SparkConf()
4848
val secret = Utils.createSecret(sparkConf)
49-
val presetPythonExec = Provenance.fromConf(sparkConf, PYSPARK_DRIVER_PYTHON)
50-
.orElse(Provenance.fromConf(sparkConf, PYSPARK_PYTHON))
49+
val presetPythonExec = Provenance.fromConfOpt(sparkConf, PYSPARK_DRIVER_PYTHON)
50+
.orElse(Provenance.fromConfOpt(sparkConf, PYSPARK_PYTHON))
5151
.orElse(Provenance.fromEnv("PYSPARK_DRIVER_PYTHON"))
5252
.orElse(Provenance.fromEnv("PYSPARK_PYTHON"))
5353

core/src/main/scala/org/apache/spark/deploy/RRunner.scala

Lines changed: 10 additions & 26 deletions
Original file line numberDiff line numberDiff line change
@@ -25,20 +25,12 @@ import scala.collection.JavaConverters._
2525

2626
import org.apache.hadoop.fs.Path
2727

28-
<<<<<<< HEAD
29-
import org.apache.spark.SparkException
30-
import org.apache.spark.SparkUserAppException
28+
import org.apache.spark.{SparkException, SparkUserAppException}
3129
import org.apache.spark.api.conda.CondaEnvironment
32-
import org.apache.spark.api.r.RBackend
33-
import org.apache.spark.api.r.RUtils
34-
import org.apache.spark.api.r.SparkRDefaults
30+
import org.apache.spark.api.r.{RBackend, RUtils}
3531
import org.apache.spark.deploy.Common.Provenance
3632
import org.apache.spark.internal.Logging
37-
=======
38-
import org.apache.spark.{SparkException, SparkUserAppException}
39-
import org.apache.spark.api.r.{RBackend, RUtils}
4033
import org.apache.spark.internal.config.R._
41-
>>>>>>> master
4234
import org.apache.spark.util.RedirectThread
4335

4436
/**
@@ -62,27 +54,19 @@ object RRunner extends CondaRunner with Logging {
6254
}
6355
// "spark.sparkr.r.command" is deprecated and replaced by "spark.r.command",
6456
// but kept here for backward compatibility.
65-
<<<<<<< HEAD
6657
driverPreset
67-
.orElse(Provenance.fromConf("spark.r.command"))
68-
.orElse(Provenance.fromConf("spark.sparkr.r.command"))
69-
=======
70-
var cmd = sys.props.getOrElse(SPARKR_COMMAND.key, SPARKR_COMMAND.defaultValue.get)
71-
cmd = sys.props.getOrElse(R_COMMAND.key, cmd)
72-
if (sys.props.getOrElse("spark.submit.deployMode", "client") == "client") {
73-
cmd = sys.props.getOrElse("spark.r.driver.command", cmd)
74-
}
75-
cmd
76-
>>>>>>> master
58+
.orElse(Provenance.fromConf(R_COMMAND.key))
59+
.orElse(Provenance.fromConf(SPARKR_COMMAND.key))
60+
.getOrElse(Provenance(SPARKR_COMMAND.key, SPARKR_COMMAND.defaultValue.get))
7761
}
7862

79-
val rCommand: String = maybeConda.map { conda =>
80-
presetRCommand.foreach { exec =>
81-
sys.error(s"It's forbidden to configure the r command when using conda, but found: $exec")
63+
val rCommand = maybeConda.map { conda =>
64+
if (presetRCommand.value != SPARKR_COMMAND.defaultValueString) {
65+
sys.error(s"It's forbidden to set the r executable " +
66+
s"when using conda, but found: ${presetRCommand.value}")
8267
}
8368
conda.condaEnvDir + "/bin/Rscript"
84-
}.orElse(presetRCommand.map(_.value))
85-
.getOrElse("Rscript")
69+
}.getOrElse(presetRCommand.value)
8670

8771
// Connection timeout set by R process on its connection to RBackend in seconds.
8872
val backendConnectionTimeout = sys.props.getOrElse(

core/src/main/scala/org/apache/spark/metrics/MetricsSystem.scala

Lines changed: 1 addition & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -98,16 +98,11 @@ private[spark] class MetricsSystem private (
9898
def start(registerStaticSources: Boolean = true) {
9999
require(!running, "Attempting to start a MetricsSystem that is already running")
100100
running = true
101-
<<<<<<< HEAD
102-
StaticSources.allSources.foreach(registerSource)
103-
registerSources()
104-
SharedMetricRegistries.getDefault.addListener(defaultListener)
105-
=======
106101
if (registerStaticSources) {
107102
StaticSources.allSources.foreach(registerSource)
108103
registerSources()
109104
}
110-
>>>>>>> master
105+
SharedMetricRegistries.getDefault.addListener(defaultListener)
111106
registerSinks()
112107
sinks.foreach(_.start())
113108
}

core/src/main/scala/org/apache/spark/scheduler/cluster/CoarseGrainedSchedulerBackend.scala

Lines changed: 2 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@ import java.util.concurrent.TimeUnit
2121
import java.util.concurrent.atomic.{AtomicInteger, AtomicReference}
2222
import javax.annotation.concurrent.GuardedBy
2323

24-
import scala.collection.mutable.{ArrayBuffer, HashMap, HashSet}
24+
import scala.collection.mutable.{HashMap, HashSet}
2525
import scala.concurrent.Future
2626

2727
import com.palantir.logsafe.SafeArg
@@ -31,13 +31,10 @@ import org.apache.hadoop.security.UserGroupInformation
3131
import org.apache.spark.{ExecutorAllocationClient, SparkEnv, SparkException, TaskState}
3232
import org.apache.spark.deploy.SparkHadoopUtil
3333
import org.apache.spark.deploy.security.HadoopDelegationTokenManager
34-
<<<<<<< HEAD
35-
import org.apache.spark.internal.SafeLogging
36-
=======
3734
import org.apache.spark.internal.Logging
35+
import org.apache.spark.internal.SafeLogging
3836
import org.apache.spark.internal.config._
3937
import org.apache.spark.internal.config.Network._
40-
>>>>>>> master
4138
import org.apache.spark.rpc._
4239
import org.apache.spark.scheduler._
4340
import org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages._
@@ -116,14 +113,9 @@ class CoarseGrainedSchedulerBackend(scheduler: TaskSchedulerImpl, val rpcEnv: Rp
116113
private val reviveThread =
117114
ThreadUtils.newDaemonSingleThreadScheduledExecutor("driver-revive-thread")
118115

119-
<<<<<<< HEAD
120-
class DriverEndpoint(override val rpcEnv: RpcEnv, sparkProperties: Seq[(String, String)])
121-
extends ThreadSafeRpcEndpoint with SafeLogging {
122-
=======
123116
class DriverEndpoint extends ThreadSafeRpcEndpoint with Logging {
124117

125118
override val rpcEnv: RpcEnv = CoarseGrainedSchedulerBackend.this.rpcEnv
126-
>>>>>>> master
127119

128120
// Executors that have been lost, but for which we don't yet know the real exit reason.
129121
protected val executorsPendingLossReason = new HashSet[String]

0 commit comments

Comments
 (0)