@@ -3,18 +3,19 @@ package datastax.astra.migrate
3
3
import com .datastax .spark .connector .cql .CassandraConnector
4
4
import org .slf4j .LoggerFactory
5
5
6
+ import org .apache .spark .SparkConf
6
7
import scala .collection .JavaConversions ._
7
8
8
9
object DiffData extends AbstractJob {
9
10
10
11
val logger = LoggerFactory .getLogger(this .getClass.getName)
11
12
logger.info(" Started Data Validation App" )
12
13
13
- diffTable(sourceConnection, destinationConnection)
14
+ diffTable(sourceConnection, destinationConnection, sc )
14
15
15
16
exitSpark
16
17
17
- private def diffTable (sourceConnection : CassandraConnector , destinationConnection : CassandraConnector ) = {
18
+ private def diffTable (sourceConnection : CassandraConnector , destinationConnection : CassandraConnector , config : SparkConf ) = {
18
19
val partitions = SplitPartitions .getRandomSubPartitions(splitSize, minPartition, maxPartition, Integer .parseInt(coveragePercent))
19
20
logger.info(" PARAM Calculated -- Total Partitions: " + partitions.size())
20
21
val parts = sContext.parallelize(partitions.toSeq, partitions.size);
@@ -23,11 +24,11 @@ object DiffData extends AbstractJob {
23
24
parts.foreach(part => {
24
25
sourceConnection.withSessionDo(sourceSession =>
25
26
destinationConnection.withSessionDo(destinationSession =>
26
- DiffJobSession .getInstance(sourceSession, destinationSession, sc )
27
+ DiffJobSession .getInstance(sourceSession, destinationSession, config )
27
28
.getDataAndDiff(part.getMin, part.getMax)))
28
29
})
29
30
30
- DiffJobSession .getInstance(null , null , sc ).printCounts(true );
31
+ DiffJobSession .getInstance(null , null , config ).printCounts(true );
31
32
}
32
33
33
34
}
0 commit comments