Skip to content

Commit d6a27d7

Browse files
authored
Merge pull request #357 from ldbc/spark3.2
Bump Spark version to 3.2
2 parents 668f1bd + b7bd556 commit d6a27d7

File tree

4 files changed

+15
-15
lines changed

4 files changed

+15
-15
lines changed

.circleci/config.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@ jobs:
2929
test:
3030
executor: my-executor
3131
environment:
32-
PLATFORM_VERSION: 2.12_spark3.1
32+
PLATFORM_VERSION: 2.12_spark3.2
3333
DATAGEN_VERSION: 0.5.0-SNAPSHOT
3434
steps:
3535
- checkout

README.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -46,30 +46,30 @@ E.g. with [pyenv](https://github.com/pyenv/pyenv) and [pyenv-virtualenv](https:/
4646
pyenv install 3.7.13
4747
pyenv virtualenv 3.7.13 ldbc_datagen_tools
4848
pyenv local ldbc_datagen_tools
49-
pip install -U pip
49+
pip install -U pip
5050
pip install ./tools
5151
```
5252
### Running locally
5353

5454
The `./tools/run.py` script is intended for **local runs**. To use it, download and extract Spark as follows.
5555

56-
#### Spark 3.1.x
56+
#### Spark 3.2.x
5757

58-
Spark 3.1.x is the recommended runtime to use. The rest of the instructions are provided assuming Spark 3.1.x.
58+
Spark 3.2.x is the recommended runtime to use. The rest of the instructions are provided assuming Spark 3.2.x.
5959

6060
To place Spark under `/opt/`:
6161

6262
```bash
63-
curl https://downloads.apache.org/spark/spark-3.1.2/spark-3.1.2-bin-hadoop3.2.tgz | sudo tar -xz -C /opt/
64-
export SPARK_HOME=/opt/spark-3.1.2-bin-hadoop3.2
63+
curl https://downloads.apache.org/spark/spark-3.2.1/spark-3.2.1-bin-hadoop3.2.tgz | sudo tar -xz -C /opt/
64+
export SPARK_HOME="/opt/spark-3.2.1-bin-hadoop3.2"
6565
export PATH="$SPARK_HOME/bin":"$PATH"
6666
```
6767

6868
To place under `~/`:
6969

7070
```bash
71-
curl https://downloads.apache.org/spark/spark-3.1.2/spark-3.1.2-bin-hadoop3.2.tgz | tar -xz -C ~/
72-
export SPARK_HOME=~/spark-3.1.2-bin-hadoop3.2
71+
curl https://downloads.apache.org/spark/spark-3.2.1/spark-3.2.1-bin-hadoop3.2.tgz | tar -xz -C ~/
72+
export SPARK_HOME=~/spark-3.2.1-bin-hadoop3.2
7373
export PATH="$SPARK_HOME/bin":"$PATH"
7474
```
7575

@@ -78,7 +78,7 @@ Both Java 8 and Java 11 are supported.
7878
Once you have Spark in place and built the JAR file, run the generator as follows:
7979

8080
```bash
81-
export PLATFORM_VERSION=2.12_spark3.1
81+
export PLATFORM_VERSION=2.12_spark3.2
8282
export DATAGEN_VERSION=0.5.0-SNAPSHOT
8383
./tools/run.py ./target/ldbc_snb_datagen_${PLATFORM_VERSION}-${DATAGEN_VERSION}.jar <runtime configuration arguments> -- <generator configuration arguments>
8484
```

pom.xml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -16,8 +16,8 @@
1616
<maven.compiler.target>${java.version}</maven.compiler.target>
1717
<scala.version>2.12.15</scala.version>
1818
<scala.compat.version>2.12</scala.compat.version>
19-
<spark.version>3.1.2</spark.version>
20-
<spark.compat.version>3.1</spark.compat.version>
19+
<spark.version>3.2.0</spark.version>
20+
<spark.compat.version>3.2</spark.compat.version>
2121
<spec2.version>4.2.0</spec2.version>
2222
</properties>
2323

@@ -294,7 +294,7 @@
294294

295295
<profiles>
296296
<profile>
297-
<id>spark3.1</id>
297+
<id>spark3.2</id>
298298
<properties>
299299
<!-- This is the default profile. -->
300300
</properties>

tools/emr/submit_datagen_job.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@
2929
'az': 'us-west-2c',
3030
'yes': False,
3131
'ec2_key': None,
32-
'emr_release': 'emr-6.3.0'
32+
'emr_release': 'emr-6.6.0'
3333
}
3434

3535
pp = pprint.PrettyPrinter(indent=2)
@@ -243,13 +243,13 @@ def submit_datagen_job(name,
243243
help='EC2 key name for SSH connection')
244244
parser.add_argument('--platform-version',
245245
default=defaults['platform_version'],
246-
help='The spark platform the JAR is compiled for formatted like {scala.compat.version}_spark{spark.compat.version}, e.g. 2.12_spark3.1')
246+
help='The spark platform the JAR is compiled for formatted like {scala.compat.version}_spark{spark.compat.version}, e.g. 2.12_spark3.2')
247247
parser.add_argument('--version',
248248
default=defaults['version'],
249249
help='LDBC SNB Datagen library version')
250250
parser.add_argument('--emr-release',
251251
default=defaults['emr_release'],
252-
help='The EMR release to use. E.g. emr-6.3.0')
252+
help='The EMR release to use. E.g. emr-6.6.0')
253253
parser.add_argument('-y', '--yes',
254254
default=defaults['yes'],
255255
action='store_true',

0 commit comments

Comments
 (0)