diff --git a/hudi-spark-datasource/hudi-spark/src/test/java/org/apache/hudi/client/TestHoodieClientMultiWriter.java b/hudi-spark-datasource/hudi-spark/src/test/java/org/apache/hudi/client/TestHoodieClientMultiWriter.java index a4e051d6f23ee..72d0db584ae8e 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/java/org/apache/hudi/client/TestHoodieClientMultiWriter.java +++ b/hudi-spark-datasource/hudi-spark/src/test/java/org/apache/hudi/client/TestHoodieClientMultiWriter.java @@ -49,6 +49,7 @@ import org.apache.hudi.common.table.view.FileSystemViewStorageType; import org.apache.hudi.common.testutils.HoodieTestTable; import org.apache.hudi.common.testutils.HoodieTestUtils; +import org.apache.hudi.common.testutils.InProcessTimeGenerator; import org.apache.hudi.common.util.CommitUtils; import org.apache.hudi.io.util.FileIOUtils; import org.apache.hudi.common.util.Option; @@ -703,19 +704,19 @@ public void testMultiWriterWithInsertsToDistinctPartitions(HoodieTableType table // Create the first commit SparkRDDWriteClient client = getHoodieWriteClient(cfg); - createCommitWithInsertsForPartition(cfg, client, "000", "001", 100, "2016/03/01"); + String firstCommitTime = InProcessTimeGenerator.createNewInstantTime(); + createCommitWithInsertsForPartition(cfg, client, "000", firstCommitTime, 100, "2016/03/01"); client.close(); int numConcurrentWriters = 5; ExecutorService executors = Executors.newFixedThreadPool(numConcurrentWriters); List> futures = new ArrayList<>(numConcurrentWriters); for (int loop = 0; loop < numConcurrentWriters; loop++) { - String newCommitTime = "00" + (loop + 2); String partition = "2016/03/0" + (loop + 2); futures.add(executors.submit(() -> { try { SparkRDDWriteClient writeClient = getHoodieWriteClient(cfg); - createCommitWithInsertsForPartition(cfg, writeClient, "001", newCommitTime, 100, partition); + createCommitWithInsertsForPartition(cfg, writeClient, "001", InProcessTimeGenerator.createNewInstantTime(), 100, partition); writeClient.close(); } catch (Exception e) { throw new RuntimeException(e);