@@ -146,6 +146,77 @@ index 698ca009b4f..57d774a3617 100644
146146
147147 -- Test tables
148148 CREATE table explain_temp1 (key int, val int) USING PARQUET;
149+ diff --git a/sql/core/src/test/resources/sql-tests/inputs/postgreSQL/int4.sql b/sql/core/src/test/resources/sql-tests/inputs/postgreSQL/int4.sql
150+ index 3a409eea348..26e9aaf215c 100644
151+ --- a/sql/core/src/test/resources/sql-tests/inputs/postgreSQL/int4.sql
152+ +++ b/sql/core/src/test/resources/sql-tests/inputs/postgreSQL/int4.sql
153+ @@ -6,6 +6,9 @@
154+ -- https://github.com/postgres/postgres/blob/REL_12_BETA2/src/test/regress/sql/int4.sql
155+ --
156+
157+ + -- TODO: https://github.com/apache/datafusion-comet/issues/551
158+ + --SET spark.comet.enabled = false
159+ +
160+ CREATE TABLE INT4_TBL(f1 int) USING parquet;
161+
162+ -- [SPARK-28023] Trim the string when cast string type to other types
163+ diff --git a/sql/core/src/test/resources/sql-tests/inputs/postgreSQL/int8.sql b/sql/core/src/test/resources/sql-tests/inputs/postgreSQL/int8.sql
164+ index fac23b4a26f..98b12ae5ccc 100644
165+ --- a/sql/core/src/test/resources/sql-tests/inputs/postgreSQL/int8.sql
166+ +++ b/sql/core/src/test/resources/sql-tests/inputs/postgreSQL/int8.sql
167+ @@ -6,6 +6,10 @@
168+ -- Test int8 64-bit integers.
169+ -- https://github.com/postgres/postgres/blob/REL_12_BETA2/src/test/regress/sql/int8.sql
170+ --
171+ +
172+ + -- TODO: https://github.com/apache/datafusion-comet/issues/551
173+ + --SET spark.comet.enabled = false
174+ +
175+ CREATE TABLE INT8_TBL(q1 bigint, q2 bigint) USING parquet;
176+
177+ -- PostgreSQL implicitly casts string literals to data with integral types, but
178+ diff --git a/sql/core/src/test/resources/sql-tests/inputs/postgreSQL/select_having.sql b/sql/core/src/test/resources/sql-tests/inputs/postgreSQL/select_having.sql
179+ index 0efe0877e9b..f9df0400c99 100644
180+ --- a/sql/core/src/test/resources/sql-tests/inputs/postgreSQL/select_having.sql
181+ +++ b/sql/core/src/test/resources/sql-tests/inputs/postgreSQL/select_having.sql
182+ @@ -6,6 +6,9 @@
183+ -- https://github.com/postgres/postgres/blob/REL_12_BETA2/src/test/regress/sql/select_having.sql
184+ --
185+
186+ + -- TODO: https://github.com/apache/datafusion-comet/issues/551
187+ + --SET spark.comet.enabled = false
188+ +
189+ -- load test data
190+ CREATE TABLE test_having (a int, b int, c string, d string) USING parquet;
191+ INSERT INTO test_having VALUES (0, 1, 'XXXX', 'A');
192+ diff --git a/sql/core/src/test/resources/sql-tests/inputs/view-schema-binding-config.sql b/sql/core/src/test/resources/sql-tests/inputs/view-schema-binding-config.sql
193+ index e803254ea64..74db78aee38 100644
194+ --- a/sql/core/src/test/resources/sql-tests/inputs/view-schema-binding-config.sql
195+ +++ b/sql/core/src/test/resources/sql-tests/inputs/view-schema-binding-config.sql
196+ @@ -1,6 +1,9 @@
197+ -- This test suits check the spark.sql.viewSchemaBindingMode configuration.
198+ -- It can be DISABLED and COMPENSATION
199+
200+ + -- TODO: https://github.com/apache/datafusion-comet/issues/551
201+ + --SET spark.comet.enabled = false
202+ +
203+ -- Verify the default binding is true
204+ SET spark.sql.legacy.viewSchemaBindingMode;
205+
206+ diff --git a/sql/core/src/test/resources/sql-tests/inputs/view-schema-compensation.sql b/sql/core/src/test/resources/sql-tests/inputs/view-schema-compensation.sql
207+ index 21a3ce1e122..f4762ab98f0 100644
208+ --- a/sql/core/src/test/resources/sql-tests/inputs/view-schema-compensation.sql
209+ +++ b/sql/core/src/test/resources/sql-tests/inputs/view-schema-compensation.sql
210+ @@ -1,5 +1,9 @@
211+ -- This test suite checks the WITH SCHEMA COMPENSATION clause
212+ -- Disable ANSI mode to ensure we are forcing it explicitly in the CASTS
213+ +
214+ + -- TODO: https://github.com/apache/datafusion-comet/issues/551
215+ + --SET spark.comet.enabled = false
216+ +
217+ SET spark.sql.ansi.enabled = false;
218+
219+ -- In COMPENSATION views get invalidated if the type can't cast
149220diff --git a/sql/core/src/test/scala/org/apache/spark/sql/CachedTableSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/CachedTableSuite.scala
150221index d023fb82185..0f4f03bda6c 100644
151222--- a/sql/core/src/test/scala/org/apache/spark/sql/CachedTableSuite.scala
@@ -917,7 +988,7 @@ index 34c6c49bc49..f5dea07a213 100644
917988 protected val baseResourcePath = {
918989 // use the same way as `SQLQueryTestSuite` to get the resource path
919990diff --git a/sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala
920- index 56c364e2084..a00a50e020a 100644
991+ index 56c364e2084..fc3abd7cdc4 100644
921992--- a/sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala
922993+++ b/sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala
923994@@ -1510,7 +1510,8 @@ class SQLQuerySuite extends QueryTest with SharedSparkSession with AdaptiveSpark
@@ -930,6 +1001,36 @@ index 56c364e2084..a00a50e020a 100644
9301001 AccumulatorSuite.verifyPeakExecutionMemorySet(sparkContext, "external sort") {
9311002 sql("SELECT * FROM testData2 ORDER BY a ASC, b ASC").collect()
9321003 }
1004+ @@ -4454,7 +4455,8 @@ class SQLQuerySuite extends QueryTest with SharedSparkSession with AdaptiveSpark
1005+ }
1006+
1007+ test("SPARK-39166: Query context of binary arithmetic should be serialized to executors" +
1008+ - " when WSCG is off") {
1009+ + " when WSCG is off",
1010+ + IgnoreComet("TODO: https://github.com/apache/datafusion-comet/issues/551")) {
1011+ withSQLConf(SQLConf.WHOLESTAGE_CODEGEN_ENABLED.key -> "false",
1012+ SQLConf.ANSI_ENABLED.key -> "true") {
1013+ withTable("t") {
1014+ @@ -4475,7 +4477,8 @@ class SQLQuerySuite extends QueryTest with SharedSparkSession with AdaptiveSpark
1015+ }
1016+
1017+ test("SPARK-39175: Query context of Cast should be serialized to executors" +
1018+ - " when WSCG is off") {
1019+ + " when WSCG is off",
1020+ + IgnoreComet("TODO: https://github.com/apache/datafusion-comet/issues/551")) {
1021+ withSQLConf(SQLConf.WHOLESTAGE_CODEGEN_ENABLED.key -> "false",
1022+ SQLConf.ANSI_ENABLED.key -> "true") {
1023+ withTable("t") {
1024+ @@ -4502,7 +4505,8 @@ class SQLQuerySuite extends QueryTest with SharedSparkSession with AdaptiveSpark
1025+ }
1026+
1027+ test("SPARK-39190,SPARK-39208,SPARK-39210: Query context of decimal overflow error should " +
1028+ - "be serialized to executors when WSCG is off") {
1029+ + "be serialized to executors when WSCG is off",
1030+ + IgnoreComet("TODO: https://github.com/apache/datafusion-comet/issues/551")) {
1031+ withSQLConf(SQLConf.WHOLESTAGE_CODEGEN_ENABLED.key -> "false",
1032+ SQLConf.ANSI_ENABLED.key -> "true") {
1033+ withTable("t") {
9331034diff --git a/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala
9341035index 68f14f13bbd..174636cefb5 100644
9351036--- a/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala
0 commit comments