@@ -58,7 +58,7 @@ protected OriginCountJobSession(CqlSession sourceSession, SparkConf sparkConf) {
58
58
filterColName = sparkConf .get ("spark.origin.FilterColumn" );
59
59
filterColType = sparkConf .get ("spark.origin.FilterColumnType" );
60
60
filterColIndex = Integer .parseInt (sparkConf .get ("spark.origin.FilterColumnIndex" , "0" ));
61
- fieldGuardraillimitMB = Integer .parseInt (sparkConf .get ("spark.fieldGuardraillimitMB" , "0" ));
61
+ fieldGuardraillimitMB = Integer .parseInt (sparkConf .get ("spark.fieldGuardraillimitMB" , "0" ));
62
62
63
63
String partionKey = sparkConf .get ("spark.query.cols.partitionKey" );
64
64
idColTypes = getTypes (sparkConf .get ("spark.query.cols.id.types" ));
@@ -94,7 +94,7 @@ public void getData(BigInteger min, BigInteger max) {
94
94
if (checkTableforColSize ) {
95
95
int rowColcnt = GetRowColumnLength (sourceRow , filterColType , filterColIndex );
96
96
String result = "" ;
97
- if (rowColcnt > fieldGuardraillimitMB ) {
97
+ if (rowColcnt > fieldGuardraillimitMB * 1048576 ) {
98
98
for (int index = 0 ; index < checkTableforColSizeTypes .size (); index ++) {
99
99
MigrateDataType dataType = checkTableforColSizeTypes .get (index );
100
100
Object colData = getData (dataType , index , sourceRow );
@@ -116,7 +116,7 @@ public void getData(BigInteger min, BigInteger max) {
116
116
if (checkTableforColSize ) {
117
117
int rowColcnt = GetRowColumnLength (sourceRow , filterColType , filterColIndex );
118
118
String result = "" ;
119
- if (rowColcnt > fieldGuardraillimitMB ) {
119
+ if (rowColcnt > fieldGuardraillimitMB * 1048576 ) {
120
120
for (int index = 0 ; index < checkTableforColSizeTypes .size (); index ++) {
121
121
MigrateDataType dataType = checkTableforColSizeTypes .get (index );
122
122
Object colData = getData (dataType , index , sourceRow );
0 commit comments