Skip to content

Commit 098c453

Browse files
committed
CDM-18 : as all defaults are now in KnownProperties, removing Utils.getSparkProp* methods in favour of PropertyHelper
1 parent 7908bbf commit 098c453

File tree

8 files changed

+97
-127
lines changed

8 files changed

+97
-127
lines changed

src/main/java/datastax/astra/migrate/AbstractJobSession.java

Lines changed: 29 additions & 42 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,6 @@
66
import com.datastax.oss.driver.api.core.cql.Row;
77
import com.datastax.oss.driver.shaded.guava.common.util.concurrent.RateLimiter;
88
import datastax.astra.migrate.properties.KnownProperties;
9-
import org.apache.commons.lang.StringUtils;
109
import org.apache.spark.SparkConf;
1110
import org.slf4j.Logger;
1211
import org.slf4j.LoggerFactory;
@@ -35,57 +34,45 @@ protected AbstractJobSession(CqlSession originSession, CqlSession targetSession,
3534
this.originSessionSession = originSession;
3635
this.targetSession = targetSession;
3736

38-
batchSize = new Integer(Util.getSparkPropOr(sc, KnownProperties.SPARK_BATCH_SIZE, "5"));
39-
fetchSizeInRows = new Integer(Util.getSparkPropOr(sc, KnownProperties.READ_FETCH_SIZE, "1000"));
40-
printStatsAfter = new Integer(Util.getSparkPropOr(sc, KnownProperties.SPARK_STATS_AFTER, "100000"));
41-
if (printStatsAfter < 1) {
42-
printStatsAfter = 100000;
37+
batchSize = propertyHelper.getInteger(KnownProperties.SPARK_BATCH_SIZE);
38+
fetchSizeInRows = propertyHelper.getInteger(KnownProperties.READ_FETCH_SIZE);
39+
printStatsAfter = propertyHelper.getInteger(KnownProperties.SPARK_STATS_AFTER);
40+
if (!propertyHelper.meetsMinimum(KnownProperties.SPARK_STATS_AFTER, printStatsAfter, 1)) {
41+
logger.warn(KnownProperties.SPARK_STATS_AFTER +" must be greater than 0. Setting to default value of " + KnownProperties.getDefaultAsString(KnownProperties.SPARK_STATS_AFTER));
42+
propertyHelper.setProperty(KnownProperties.SPARK_STATS_AFTER, KnownProperties.getDefault(KnownProperties.SPARK_STATS_AFTER));
43+
printStatsAfter = propertyHelper.getInteger(KnownProperties.SPARK_STATS_AFTER);
4344
}
4445

45-
readLimiter = RateLimiter.create(new Integer(Util.getSparkPropOr(sc, KnownProperties.SPARK_LIMIT_READ, "20000")));
46-
writeLimiter = RateLimiter.create(new Integer(Util.getSparkPropOr(sc, KnownProperties.SPARK_LIMIT_WRITE, "40000")));
47-
maxRetries = Integer.parseInt(sc.get(KnownProperties.SPARK_MAX_RETRIES, "0"));
46+
readLimiter = RateLimiter.create(propertyHelper.getInteger(KnownProperties.SPARK_LIMIT_READ));
47+
writeLimiter = RateLimiter.create(propertyHelper.getInteger(KnownProperties.SPARK_LIMIT_WRITE));
48+
maxRetries = propertyHelper.getInteger(KnownProperties.SPARK_MAX_RETRIES);
4849

49-
originKeyspaceTable = Util.getSparkProp(sc, KnownProperties.ORIGIN_KEYSPACE_TABLE);
50-
targetKeyspaceTable = Util.getSparkProp(sc, KnownProperties.TARGET_KEYSPACE_TABLE);
50+
originKeyspaceTable = propertyHelper.getString(KnownProperties.ORIGIN_KEYSPACE_TABLE);
51+
targetKeyspaceTable = propertyHelper.getString(KnownProperties.TARGET_KEYSPACE_TABLE);
5152

52-
String ttlColsStr = Util.getSparkPropOrEmpty(sc, KnownProperties.ORIGIN_TTL_COLS);
53+
String ttlColsStr = propertyHelper.getAsString(KnownProperties.ORIGIN_TTL_COLS);
5354
if (null != ttlColsStr && ttlColsStr.trim().length() > 0) {
5455
for (String ttlCol : ttlColsStr.split(",")) {
5556
ttlCols.add(Integer.parseInt(ttlCol));
5657
}
5758
}
5859

59-
String writeTimestampColsStr = Util.getSparkPropOrEmpty(sc, KnownProperties.ORIGIN_WRITETIME_COLS);
60+
String writeTimestampColsStr = propertyHelper.getAsString(KnownProperties.ORIGIN_WRITETIME_COLS);
6061
if (null != writeTimestampColsStr && writeTimestampColsStr.trim().length() > 0) {
6162
for (String writeTimeStampCol : writeTimestampColsStr.split(",")) {
6263
writeTimeStampCols.add(Integer.parseInt(writeTimeStampCol));
6364
}
6465
}
6566

66-
writeTimeStampFilter = Boolean
67-
.parseBoolean(Util.getSparkPropOr(sc, KnownProperties.ORIGIN_FILTER_WRITETS_ENABLED, "false"));
67+
writeTimeStampFilter = propertyHelper.getBoolean(KnownProperties.ORIGIN_FILTER_WRITETS_ENABLED);
6868
// batchsize set to 1 if there is a writeFilter
6969
if (writeTimeStampFilter) {
7070
batchSize = 1;
7171
}
7272

73-
String minWriteTimeStampFilterStr =
74-
Util.getSparkPropOr(sc, KnownProperties.ORIGIN_FILTER_WRITETS_MIN, "0");
75-
if (null != minWriteTimeStampFilterStr && minWriteTimeStampFilterStr.trim().length() > 1) {
76-
minWriteTimeStampFilter = Long.parseLong(minWriteTimeStampFilterStr);
77-
}
78-
String maxWriteTimeStampFilterStr =
79-
Util.getSparkPropOr(sc, KnownProperties.ORIGIN_FILTER_WRITETS_MAX, "0");
80-
if (null != maxWriteTimeStampFilterStr && maxWriteTimeStampFilterStr.trim().length() > 1) {
81-
maxWriteTimeStampFilter = Long.parseLong(maxWriteTimeStampFilterStr);
82-
}
83-
84-
String customWriteTimeStr =
85-
Util.getSparkPropOr(sc, KnownProperties.TARGET_CUSTOM_WRITETIME, "0");
86-
if (null != customWriteTimeStr && customWriteTimeStr.trim().length() > 1 && StringUtils.isNumeric(customWriteTimeStr.trim())) {
87-
customWritetime = Long.parseLong(customWriteTimeStr);
88-
}
73+
minWriteTimeStampFilter = propertyHelper.getLong(KnownProperties.ORIGIN_FILTER_WRITETS_MIN);
74+
maxWriteTimeStampFilter = propertyHelper.getLong(KnownProperties.ORIGIN_FILTER_WRITETS_MAX);
75+
customWritetime = propertyHelper.getLong(KnownProperties.TARGET_CUSTOM_WRITETIME);
8976

9077
logger.info("PARAM -- Read Consistency: {}", readConsistencyLevel);
9178
logger.info("PARAM -- Write Consistency: {}", writeConsistencyLevel);
@@ -106,9 +93,9 @@ protected AbstractJobSession(CqlSession originSession, CqlSession targetSession,
10693
Instant.ofEpochMilli(maxWriteTimeStampFilter / 1000));
10794
}
10895

109-
String selectCols = Util.getSparkProp(sc, KnownProperties.ORIGIN_COLUMN_NAMES);
110-
String partitionKey = Util.getSparkProp(sc, KnownProperties.ORIGIN_PARTITION_KEY);
111-
String originSelectCondition = Util.getSparkPropOrEmpty(sc, KnownProperties.ORIGIN_FILTER_CONDITION);
96+
String selectCols = propertyHelper.getAsString(KnownProperties.ORIGIN_COLUMN_NAMES);
97+
String partitionKey = propertyHelper.getAsString(KnownProperties.ORIGIN_PARTITION_KEY);
98+
String originSelectCondition = propertyHelper.getAsString(KnownProperties.ORIGIN_FILTER_CONDITION);
11299
if (!originSelectCondition.isEmpty() && !originSelectCondition.trim().toUpperCase().startsWith("AND")) {
113100
originSelectCondition = " AND " + originSelectCondition;
114101
}
@@ -121,11 +108,11 @@ protected AbstractJobSession(CqlSession originSession, CqlSession targetSession,
121108
writeTimeStampCols.forEach(col -> {
122109
selectTTLWriteTimeCols.append(",writetime(" + allCols[col] + ")");
123110
});
124-
selectColTypes = getTypes(Util.getSparkProp(sc, KnownProperties.ORIGIN_COLUMN_TYPES));
125-
String idCols = Util.getSparkPropOrEmpty(sc, KnownProperties.TARGET_PRIMARY_KEY);
111+
selectColTypes = getTypes(propertyHelper.getAsString(KnownProperties.ORIGIN_COLUMN_TYPES));
112+
String idCols = propertyHelper.getAsString(KnownProperties.TARGET_PRIMARY_KEY);
126113
idColTypes = selectColTypes.subList(0, idCols.split(",").length);
127114

128-
String insertCols = Util.getSparkPropOrEmpty(sc, KnownProperties.TARGET_COLUMN_NAMES);
115+
String insertCols = propertyHelper.getAsString(KnownProperties.TARGET_COLUMN_NAMES);
129116
if (null == insertCols || insertCols.trim().isEmpty()) {
130117
insertCols = selectCols;
131118
}
@@ -153,15 +140,15 @@ protected AbstractJobSession(CqlSession originSession, CqlSession targetSession,
153140
"select " + insertCols + " from " + targetKeyspaceTable
154141
+ " where " + insertBinds);
155142

156-
hasRandomPartitioner = Boolean.parseBoolean(Util.getSparkPropOr(sc, KnownProperties.ORIGIN_HAS_RANDOM_PARTITIONER, "false"));
157-
isCounterTable = Boolean.parseBoolean(Util.getSparkPropOr(sc, KnownProperties.ORIGIN_IS_COUNTER, "false"));
143+
hasRandomPartitioner = propertyHelper.getBoolean(KnownProperties.ORIGIN_HAS_RANDOM_PARTITIONER);
144+
isCounterTable = propertyHelper.getBoolean(KnownProperties.ORIGIN_IS_COUNTER);
158145
if (isCounterTable) {
159-
String updateSelectMappingStr = Util.getSparkPropOr(sc, KnownProperties.ORIGIN_COUNTER_INDEXES, "0");
146+
String updateSelectMappingStr = propertyHelper.getString(KnownProperties.ORIGIN_COUNTER_INDEXES);
160147
for (String updateSelectIndex : updateSelectMappingStr.split(",")) {
161148
updateSelectMapping.add(Integer.parseInt(updateSelectIndex));
162149
}
163150

164-
String counterTableUpdate = Util.getSparkProp(sc, KnownProperties.ORIGIN_COUNTER_CQL);
151+
String counterTableUpdate = propertyHelper.getString(KnownProperties.ORIGIN_COUNTER_CQL);
165152
targetInsertStatement = targetSession.prepare(counterTableUpdate);
166153
} else {
167154
insertBinds = "";
@@ -186,7 +173,7 @@ protected AbstractJobSession(CqlSession originSession, CqlSession targetSession,
186173
}
187174

188175
// Handle rows with blank values for 'timestamp' data-type in primary-key fields
189-
tsReplaceValStr = Util.getSparkPropOr(sc, KnownProperties.TARGET_REPLACE_MISSING_TS, "");
176+
tsReplaceValStr = propertyHelper.getAsString(KnownProperties.TARGET_REPLACE_MISSING_TS);
190177
if (!tsReplaceValStr.isEmpty()) {
191178
tsReplaceVal = Long.parseLong(tsReplaceValStr);
192179
}

src/main/java/datastax/astra/migrate/BaseJobSession.java

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,6 +6,7 @@
66
import com.datastax.oss.driver.api.core.cql.Row;
77
import com.datastax.oss.driver.shaded.guava.common.util.concurrent.RateLimiter;
88
import datastax.astra.migrate.properties.KnownProperties;
9+
import datastax.astra.migrate.properties.PropertyHelper;
910
import org.apache.spark.SparkConf;
1011

1112
import java.util.ArrayList;
@@ -15,6 +16,8 @@
1516

1617
public abstract class BaseJobSession {
1718

19+
protected PropertyHelper propertyHelper = PropertyHelper.getInstance();
20+
1821
protected PreparedStatement originSelectStatement;
1922
protected PreparedStatement targetSelectStatement;
2023
protected PreparedStatement targetInsertStatement;
@@ -65,8 +68,9 @@ public abstract class BaseJobSession {
6568
protected long tsReplaceVal;
6669

6770
protected BaseJobSession(SparkConf sc) {
68-
readConsistencyLevel = Util.mapToConsistencyLevel(Util.getSparkPropOrEmpty(sc, KnownProperties.READ_CL));
69-
writeConsistencyLevel = Util.mapToConsistencyLevel(Util.getSparkPropOrEmpty(sc, KnownProperties.WRITE_CL));
71+
propertyHelper.initializeSparkConf(sc);
72+
readConsistencyLevel = Util.mapToConsistencyLevel(propertyHelper.getString(KnownProperties.READ_CL));
73+
writeConsistencyLevel = Util.mapToConsistencyLevel(propertyHelper.getString(KnownProperties.WRITE_CL));
7074
}
7175

7276
public String getKey(Row originRow) {

src/main/java/datastax/astra/migrate/CopyJobSession.java

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -24,11 +24,11 @@ public class CopyJobSession extends AbstractJobSession {
2424

2525
protected CopyJobSession(CqlSession originSession, CqlSession targetSession, SparkConf sc) {
2626
super(originSession, targetSession, sc);
27-
filterData = Boolean.parseBoolean(sc.get(KnownProperties.ORIGIN_FILTER_COLUMN_ENABLED, "false"));
28-
filterColName = Util.getSparkPropOrEmpty(sc, KnownProperties.ORIGIN_FILTER_COLUMN_NAME);
29-
filterColType = Util.getSparkPropOrEmpty(sc, KnownProperties.ORIGIN_FILTER_COLUMN_TYPE);
30-
filterColIndex = Integer.parseInt(sc.get(KnownProperties.ORIGIN_FILTER_COLUMN_INDEX, "0"));
31-
filterColValue = Util.getSparkPropOrEmpty(sc, KnownProperties.ORIGIN_FILTER_COLUMN_VALUE);
27+
filterData = propertyHelper.getBoolean(KnownProperties.ORIGIN_FILTER_COLUMN_ENABLED);
28+
filterColName = propertyHelper.getAsString(KnownProperties.ORIGIN_FILTER_COLUMN_NAME);
29+
filterColType = propertyHelper.getAsString(KnownProperties.ORIGIN_FILTER_COLUMN_TYPE);
30+
filterColIndex = propertyHelper.getInteger(KnownProperties.ORIGIN_FILTER_COLUMN_INDEX);
31+
filterColValue = propertyHelper.getAsString(KnownProperties.ORIGIN_FILTER_COLUMN_VALUE);
3232
}
3333

3434
public static CopyJobSession getInstance(CqlSession originSession, CqlSession targetSession, SparkConf sc) {

src/main/java/datastax/astra/migrate/DiffJobSession.java

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -37,10 +37,10 @@ public class DiffJobSession extends CopyJobSession {
3737
private DiffJobSession(CqlSession originSession, CqlSession targetSession, SparkConf sc) {
3838
super(originSession, targetSession, sc);
3939

40-
autoCorrectMissing = Boolean.parseBoolean(Util.getSparkPropOr(sc, KnownProperties.TARGET_AUTOCORRECT_MISSING, "false"));
40+
autoCorrectMissing = propertyHelper.getBoolean(KnownProperties.TARGET_AUTOCORRECT_MISSING);
4141
logger.info("PARAM -- Autocorrect Missing: {}", autoCorrectMissing);
4242

43-
autoCorrectMismatch = Boolean.parseBoolean(Util.getSparkPropOr(sc, KnownProperties.TARGET_AUTOCORRECT_MISMATCH, "false"));
43+
autoCorrectMismatch = propertyHelper.getBoolean(KnownProperties.TARGET_AUTOCORRECT_MISMATCH);
4444
logger.info("PARAM -- Autocorrect Mismatch: {}", autoCorrectMismatch);
4545
}
4646

src/main/java/datastax/astra/migrate/OriginCountJobSession.java

Lines changed: 22 additions & 20 deletions
Original file line numberDiff line numberDiff line change
@@ -30,35 +30,37 @@ public class OriginCountJobSession extends BaseJobSession {
3030
protected OriginCountJobSession(CqlSession originSession, SparkConf sc) {
3131
super(sc);
3232
this.originSessionSession = originSession;
33-
batchSize = new Integer(sc.get(KnownProperties.SPARK_BATCH_SIZE, "1"));
34-
printStatsAfter = new Integer(sc.get(KnownProperties.SPARK_STATS_AFTER, "100000"));
35-
if (printStatsAfter < 1) {
36-
printStatsAfter = 100000;
33+
batchSize = propertyHelper.getInteger(KnownProperties.SPARK_BATCH_SIZE);
34+
printStatsAfter = propertyHelper.getInteger(KnownProperties.SPARK_STATS_AFTER);
35+
if (!propertyHelper.meetsMinimum(KnownProperties.SPARK_STATS_AFTER, printStatsAfter, 1)) {
36+
logger.warn(KnownProperties.SPARK_STATS_AFTER +" must be greater than 0. Setting to default value of " + KnownProperties.getDefaultAsString(KnownProperties.SPARK_STATS_AFTER));
37+
propertyHelper.setProperty(KnownProperties.SPARK_STATS_AFTER, KnownProperties.getDefault(KnownProperties.SPARK_STATS_AFTER));
38+
printStatsAfter = propertyHelper.getInteger(KnownProperties.SPARK_STATS_AFTER);
3739
}
3840

39-
readLimiter = RateLimiter.create(new Integer(sc.get(KnownProperties.SPARK_LIMIT_READ, "20000")));
40-
originKeyspaceTable = sc.get(KnownProperties.ORIGIN_KEYSPACE_TABLE);
41+
readLimiter = RateLimiter.create(propertyHelper.getInteger(KnownProperties.SPARK_LIMIT_READ));
42+
originKeyspaceTable = propertyHelper.getString(KnownProperties.ORIGIN_KEYSPACE_TABLE);
4143

42-
hasRandomPartitioner = Boolean.parseBoolean(sc.get(KnownProperties.ORIGIN_HAS_RANDOM_PARTITIONER, "false"));
43-
isCounterTable = Boolean.parseBoolean(sc.get(KnownProperties.ORIGIN_IS_COUNTER, "false"));
44+
hasRandomPartitioner = propertyHelper.getBoolean(KnownProperties.ORIGIN_HAS_RANDOM_PARTITIONER);
45+
isCounterTable = propertyHelper.getBoolean(KnownProperties.ORIGIN_IS_COUNTER);
4446

45-
checkTableforColSize = Boolean.parseBoolean(sc.get(KnownProperties.ORIGIN_CHECK_COLSIZE_ENABLED, "false"));
46-
checkTableforselectCols = sc.get(KnownProperties.ORIGIN_CHECK_COLSIZE_COLUMN_NAMES);
47-
checkTableforColSizeTypes = getTypes(sc.get(KnownProperties.ORIGIN_CHECK_COLSIZE_COLUMN_TYPES));
48-
filterColName = Util.getSparkPropOrEmpty(sc, KnownProperties.ORIGIN_FILTER_COLUMN_NAME);
49-
filterColType = Util.getSparkPropOrEmpty(sc, KnownProperties.ORIGIN_FILTER_COLUMN_TYPE);
50-
filterColIndex = Integer.parseInt(sc.get(KnownProperties.ORIGIN_FILTER_COLUMN_INDEX, "0"));
51-
fieldGuardraillimitMB = Integer.parseInt(sc.get(KnownProperties.FIELD_GUARDRAIL_MB, "0"));
47+
checkTableforColSize = propertyHelper.getBoolean(KnownProperties.ORIGIN_CHECK_COLSIZE_ENABLED);
48+
checkTableforselectCols = propertyHelper.getAsString(KnownProperties.ORIGIN_CHECK_COLSIZE_COLUMN_NAMES);
49+
checkTableforColSizeTypes = getTypes(propertyHelper.getAsString(KnownProperties.ORIGIN_CHECK_COLSIZE_COLUMN_TYPES));
50+
filterColName = propertyHelper.getAsString(KnownProperties.ORIGIN_FILTER_COLUMN_NAME);
51+
filterColType = propertyHelper.getAsString(KnownProperties.ORIGIN_FILTER_COLUMN_TYPE);
52+
filterColIndex = propertyHelper.getInteger(KnownProperties.ORIGIN_FILTER_COLUMN_INDEX);
53+
fieldGuardraillimitMB = propertyHelper.getInteger(KnownProperties.FIELD_GUARDRAIL_MB);
5254

53-
String partionKey = sc.get(KnownProperties.ORIGIN_PARTITION_KEY);
54-
idColTypes = getTypes(sc.get(KnownProperties.TARGET_PRIMARY_KEY_TYPES));
55+
String partionKey = propertyHelper.getAsString(KnownProperties.ORIGIN_PARTITION_KEY);
56+
idColTypes = getTypes(propertyHelper.getAsString(KnownProperties.TARGET_PRIMARY_KEY_TYPES));
5557

56-
String selectCols = sc.get(KnownProperties.ORIGIN_COLUMN_NAMES);
57-
String updateSelectMappingStr = sc.get(KnownProperties.ORIGIN_COUNTER_INDEXES, "0");
58+
String selectCols = propertyHelper.getAsString(KnownProperties.ORIGIN_COLUMN_NAMES);
59+
String updateSelectMappingStr = propertyHelper.getAsString(KnownProperties.ORIGIN_COUNTER_INDEXES);
5860
for (String updateSelectIndex : updateSelectMappingStr.split(",")) {
5961
updateSelectMapping.add(Integer.parseInt(updateSelectIndex));
6062
}
61-
String originSelectCondition = sc.get(KnownProperties.ORIGIN_FILTER_CONDITION, "");
63+
String originSelectCondition = propertyHelper.getAsString(KnownProperties.ORIGIN_FILTER_CONDITION);
6264
// TODO: AbstractJobSession has some checks to ensure AND is added to the condition
6365
originSelectStatement = originSession.prepare(
6466
"select " + selectCols + " from " + originKeyspaceTable + " where token(" + partionKey.trim()

src/main/java/datastax/astra/migrate/Util.java

Lines changed: 0 additions & 24 deletions
Original file line numberDiff line numberDiff line change
@@ -1,38 +1,14 @@
11
package datastax.astra.migrate;
22

33
import com.datastax.oss.driver.api.core.ConsistencyLevel;
4-
import datastax.astra.migrate.properties.KnownProperties;
5-
import datastax.astra.migrate.properties.PropertyHelper;
64
import org.apache.commons.lang.StringUtils;
7-
import org.apache.spark.SparkConf;
85

96
import java.io.BufferedReader;
107
import java.io.FileNotFoundException;
118
import java.io.FileReader;
12-
import java.util.NoSuchElementException;
139

1410
public class Util {
1511

16-
public static String getSparkProp(SparkConf sc, String prop) {
17-
String retVal = PropertyHelper.getInstance(sc).getAsString(prop);
18-
if (StringUtils.isEmpty(retVal) && (prop.contains("origin") || prop.contains("target"))) {
19-
retVal = PropertyHelper.getInstance(sc).getAsString(prop.replace("origin", "source").replace("target", "destination"));
20-
}
21-
if (!KnownProperties.isKnown(prop)) {
22-
throw new IllegalArgumentException("Unknown property: " + prop + "; this is a bug in the code: the property is not configured in KnownProperties.java");
23-
}
24-
return retVal;
25-
}
26-
27-
public static String getSparkPropOr(SparkConf sc, String prop, String defaultVal) {
28-
String retVal = getSparkProp(sc,prop);
29-
return StringUtils.isEmpty(retVal) ? defaultVal : retVal;
30-
}
31-
32-
public static String getSparkPropOrEmpty(SparkConf sc, String prop) {
33-
return getSparkPropOr(sc, prop, "");
34-
}
35-
3612
public static BufferedReader getfileReader(String fileName) {
3713
try {
3814
return new BufferedReader(new FileReader(fileName));

src/main/java/datastax/astra/migrate/properties/KnownProperties.java

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -196,7 +196,7 @@ public enum PropertyType {
196196
types.put(SPARK_STATS_AFTER, PropertyType.NUMBER);
197197
defaults.put(SPARK_STATS_AFTER, "100000");
198198
types.put(FIELD_GUARDRAIL_MB, PropertyType.NUMBER);
199-
defaults.put(FIELD_GUARDRAIL_MB, "10");
199+
defaults.put(FIELD_GUARDRAIL_MB, "0");
200200
types.put(PARTITION_MIN, PropertyType.NUMBER);
201201
defaults.put(PARTITION_MIN, "-9223372036854775808");
202202
types.put(PARTITION_MAX, PropertyType.NUMBER);

0 commit comments

Comments
 (0)