@@ -105,6 +105,7 @@ def test_partitioned_table_rewrite(spark: SparkSession, session_catalog: RestCat
105105 assert tbl .scan ().to_arrow ().to_pydict () == {"number_partitioned" : [11 , 10 ], "number" : [30 , 30 ]}
106106
107107
108+ @pytest .mark .integration
108109@pytest .mark .parametrize ("format_version" , [1 , 2 ])
109110def test_rewrite_partitioned_table_with_null (spark : SparkSession , session_catalog : RestCatalog , format_version : int ) -> None :
110111 identifier = "default.table_partitioned_delete"
@@ -453,6 +454,7 @@ def test_delete_truncate(session_catalog: RestCatalog) -> None:
453454 assert entries [0 ].status == ManifestEntryStatus .DELETED
454455
455456
457+ @pytest .mark .integration
456458def test_delete_overwrite_table_with_null (session_catalog : RestCatalog ) -> None :
457459 arrow_schema = pa .schema ([pa .field ("ints" , pa .int32 ())])
458460 arrow_tbl = pa .Table .from_pylist (
@@ -492,6 +494,7 @@ def test_delete_overwrite_table_with_null(session_catalog: RestCatalog) -> None:
492494 assert tbl .scan ().to_arrow ()["ints" ].to_pylist () == [3 , 4 , 1 , None ]
493495
494496
497+ @pytest .mark .integration
495498def test_delete_overwrite_table_with_nan (session_catalog : RestCatalog ) -> None :
496499 arrow_schema = pa .schema ([pa .field ("floats" , pa .float32 ())])
497500
0 commit comments