Skip to content

Commit f98c0ad

Browse files
committed
[SPARK-25906][SHELL] Documents '-I' option (from Scala REPL) in spark-shell
## What changes were proposed in this pull request? This PR targets to document `-I` option from Spark 2.4.x (previously `-i` option until Spark 2.3.x). After we upgraded Scala to 2.11.12, `-i` option (`:load`) was replaced to `-I`(SI-7898). Existing `-i` became `:paste` which does not respect Spark's implicit import (for instance `toDF`, symbol as column, etc.). Therefore, `-i` option does not correctly from Spark 2.4.x and it's not documented. I checked other Scala REPL options but looks not applicable or working from quick tests. This PR only targets to document `-I` for now. ## How was this patch tested? Manually tested. **Mac:** ```bash $ ./bin/spark-shell --help Usage: ./bin/spark-shell [options] Scala REPL options: -I <file> preload <file>, enforcing line-by-line interpretation Options: --master MASTER_URL spark://host:port, mesos://host:port, yarn, k8s://https://host:port, or local (Default: local[*]). --deploy-mode DEPLOY_MODE Whether to launch the driver program locally ("client") or on one of the worker machines inside the cluster ("cluster") (Default: client). ... ``` **Windows:** ```cmd C:\...\spark>.\bin\spark-shell --help Usage: .\bin\spark-shell.cmd [options] Scala REPL options: -I <file> preload <file>, enforcing line-by-line interpretation Options: --master MASTER_URL spark://host:port, mesos://host:port, yarn, k8s://https://host:port, or local (Default: local[*]). --deploy-mode DEPLOY_MODE Whether to launch the driver program locally ("client") or on one of the worker machines inside the cluster ("cluster") (Default: client). ... ``` Closes apache#22919 from HyukjinKwon/SPARK-25906. Authored-by: hyukjinkwon <[email protected]> Signed-off-by: hyukjinkwon <[email protected]> (cherry picked from commit cc38abc) Signed-off-by: hyukjinkwon <[email protected]>
1 parent 8526f2e commit f98c0ad

File tree

2 files changed

+11
-2
lines changed

2 files changed

+11
-2
lines changed

bin/spark-shell

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -32,7 +32,10 @@ if [ -z "${SPARK_HOME}" ]; then
3232
source "$(dirname "$0")"/find-spark-home
3333
fi
3434

35-
export _SPARK_CMD_USAGE="Usage: ./bin/spark-shell [options]"
35+
export _SPARK_CMD_USAGE="Usage: ./bin/spark-shell [options]
36+
37+
Scala REPL options:
38+
-I <file> preload <file>, enforcing line-by-line interpretation"
3639

3740
# SPARK-4161: scala does not assume use of the java classpath,
3841
# so we need to add the "-Dscala.usejavacp=true" flag manually. We

bin/spark-shell2.cmd

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,13 @@ rem
2020
rem Figure out where the Spark framework is installed
2121
call "%~dp0find-spark-home.cmd"
2222

23-
set _SPARK_CMD_USAGE=Usage: .\bin\spark-shell.cmd [options]
23+
set LF=^
24+
25+
26+
rem two empty lines are required
27+
set _SPARK_CMD_USAGE=Usage: .\bin\spark-shell.cmd [options]^%LF%%LF%^%LF%%LF%^
28+
Scala REPL options:^%LF%%LF%^
29+
-I ^<file^> preload ^<file^>, enforcing line-by-line interpretation
2430

2531
rem SPARK-4161: scala does not assume use of the java classpath,
2632
rem so we need to add the "-Dscala.usejavacp=true" flag manually. We

0 commit comments

Comments
 (0)