File tree Expand file tree Collapse file tree 2 files changed +19
-3
lines changed
job-server/src/spark.jobserver Expand file tree Collapse file tree 2 files changed +19
-3
lines changed Original file line number Diff line number Diff line change 1
1
#! /bin/bash
2
2
# Script to start the job manager
3
- # args: <work dir for context> <cluster address>
3
+ # args: <work dir for context> <cluster address> [proxy_user]
4
4
set -e
5
5
6
6
get_abs_script_path () {
@@ -27,10 +27,18 @@ JAVA_OPTS="-XX:MaxDirectMemorySize=$MAX_DIRECT_MEMORY
27
27
28
28
MAIN=" spark.jobserver.JobManager"
29
29
30
- cmd=' $SPARK_HOME/bin/spark-submit --class $MAIN --driver-memory $JOBSERVER_MEMORY
30
+ if [ ! -z $3 ]; then
31
+ cmd=' $SPARK_HOME/bin/spark-submit --class $MAIN --driver-memory $JOBSERVER_MEMORY
31
32
--conf "spark.executor.extraJavaOptions=$LOGGING_OPTS"
33
+ --proxy-user $3
32
34
--driver-java-options "$GC_OPTS $JAVA_OPTS $LOGGING_OPTS $CONFIG_OVERRIDES"
33
35
$appdir/spark-job-server.jar $1 $2 $conffile'
36
+ else
37
+ cmd=' $SPARK_HOME/bin/spark-submit --class $MAIN --driver-memory $JOBSERVER_MEMORY
38
+ --conf "spark.executor.extraJavaOptions=$LOGGING_OPTS"
39
+ --driver-java-options "$GC_OPTS $JAVA_OPTS $LOGGING_OPTS $CONFIG_OVERRIDES"
40
+ $appdir/spark-job-server.jar $1 $2 $conffile'
41
+ fi
34
42
35
43
eval $cmd > /dev/null 2>&1 &
36
44
# exec java -cp $CLASSPATH $GC_OPTS $JAVA_OPTS $LOGGING_OPTS $CONFIG_OVERRIDES $MAIN $1 $2 $conffile 2>&1 &
Original file line number Diff line number Diff line change @@ -201,7 +201,15 @@ class AkkaClusterSupervisorActor(daoActor: ActorRef) extends InstrumentedActor {
201
201
failureFunc(e)
202
202
return
203
203
}
204
- val pb = Process (s " $managerStartCommand $contextDir ${selfAddress.toString}" )
204
+
205
+ // extract spark.proxy.user from contextConfig, if available and pass it to $managerStartCommand
206
+ var cmdString = s " $managerStartCommand $contextDir ${selfAddress.toString}"
207
+
208
+ if (contextConfig.hasPath(" spark.proxy.user" )) {
209
+ cmdString = cmdString + s " ${contextConfig.getString(" spark.proxy.user" )}"
210
+ }
211
+
212
+ val pb = Process (cmdString)
205
213
val pio = new ProcessIO (_ => (),
206
214
stdout => scala.io.Source .fromInputStream(stdout)
207
215
.getLines.foreach(println),
You can’t perform that action at this time.
0 commit comments