Skip to content

Commit 80c254e

Browse files
authored
Minor version bumps (#290)
* Cassandra bump to 5.0-rc1 * Bump Spark to 3.5.2
1 parent 4280a78 commit 80c254e

File tree

3 files changed

+8
-8
lines changed

3 files changed

+8
-8
lines changed

Dockerfile

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -9,9 +9,9 @@ RUN mkdir -p /assets/ && cd /assets && \
99
curl -OL https://downloads.datastax.com/enterprise/cqlsh-astra.tar.gz && \
1010
tar -xzf ./cqlsh-astra.tar.gz && \
1111
rm ./cqlsh-astra.tar.gz && \
12-
curl -OL https://archive.apache.org/dist/spark/spark-3.5.1/spark-3.5.1-bin-hadoop3-scala2.13.tgz && \
13-
tar -xzf ./spark-3.5.1-bin-hadoop3-scala2.13.tgz && \
14-
rm ./spark-3.5.1-bin-hadoop3-scala2.13.tgz
12+
curl -OL https://archive.apache.org/dist/spark/spark-3.5.2/spark-3.5.2-bin-hadoop3-scala2.13.tgz && \
13+
tar -xzf ./spark-3.5.2-bin-hadoop3-scala2.13.tgz && \
14+
rm ./spark-3.5.2-bin-hadoop3-scala2.13.tgz
1515

1616
RUN apt-get update && apt-get install -y openssh-server vim python3 --no-install-recommends && \
1717
rm -rf /var/lib/apt/lists/* && \
@@ -44,7 +44,7 @@ RUN chmod +x ./get-latest-maven-version.sh && \
4444
rm -rf "$USER_HOME_DIR/.m2"
4545

4646
# Add all migration tools to path
47-
ENV PATH="${PATH}:/assets/dsbulk/bin/:/assets/cqlsh-astra/bin/:/assets/spark-3.5.1-bin-hadoop3-scala2.13/bin/"
47+
ENV PATH="${PATH}:/assets/dsbulk/bin/:/assets/cqlsh-astra/bin/:/assets/spark-3.5.2-bin-hadoop3-scala2.13/bin/"
4848

4949
EXPOSE 22
5050

README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -18,10 +18,10 @@ Migrate and Validate Tables between Origin and Target Cassandra Clusters.
1818

1919
### Prerequisite
2020
- Install **Java11** (minimum) as Spark binaries are compiled with it.
21-
- Install Spark version [`3.5.1`](https://archive.apache.org/dist/spark/spark-3.5.1/spark-3.5.1-bin-hadoop3-scala2.13.tgz) on a single VM (no cluster necessary) where you want to run this job. Spark can be installed by running the following: -
21+
- Install Spark version [`3.5.2`](https://archive.apache.org/dist/spark/spark-3.5.2/spark-3.5.2-bin-hadoop3-scala2.13.tgz) on a single VM (no cluster necessary) where you want to run this job. Spark can be installed by running the following: -
2222
```
23-
wget https://archive.apache.org/dist/spark/spark-3.5.1/spark-3.5.1-bin-hadoop3-scala2.13.tgz
24-
tar -xvzf spark-3.5.1-bin-hadoop3-scala2.13.tgz
23+
wget https://archive.apache.org/dist/spark/spark-3.5.2/spark-3.5.2-bin-hadoop3-scala2.13.tgz
24+
tar -xvzf spark-3.5.2-bin-hadoop3-scala2.13.tgz
2525
```
2626

2727
> :warning: If the above Spark and Scala version is not properly installed, you'll then see a similar exception like below when running the CDM jobs,

pom.xml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@
1414
<scala.main.version>2.13</scala.main.version>
1515
<spark.version>3.5.2</spark.version>
1616
<connector.version>3.5.1</connector.version>
17-
<cassandra.version>5.0-beta1</cassandra.version>
17+
<cassandra.version>5.0-rc1</cassandra.version>
1818
<junit.version>5.9.1</junit.version>
1919
<mockito.version>4.11.0</mockito.version>
2020
<java-driver.version>4.18.1</java-driver.version>

0 commit comments

Comments
 (0)