@@ -3452,24 +3452,12 @@ def execute_spark_query(query: str):
34523452 raise
34533453
34543454
3455- < << << << HEAD
34563455@pytest .mark .parametrize ("storage_type" , ["s3" , "azure" ])
34573456@pytest .mark .parametrize ("run_on_cluster" , [False , True ])
34583457def test_read_constant_columns_optimization (started_cluster , storage_type , run_on_cluster ):
34593458 instance = started_cluster .instances ["node1" ]
34603459 spark = started_cluster .spark_session
34613460 TABLE_NAME = "test_read_constant_columns_optimization_" + storage_type + "_" + get_uuid_str ()
3462- == == == =
3463- @pytest .mark .parametrize (
3464- "storage_type" ,
3465- ["s3" , "azure" , "local" ],
3466- )
3467- def test_partition_pruning_with_subquery_set (started_cluster , storage_type ):
3468- instance = started_cluster .instances ["node1" ]
3469- spark = started_cluster .spark_session
3470- TABLE_NAME = "test_partition_pruning_" + storage_type + "_" + get_uuid_str ()
3471- IN_MEMORY_TABLE = "in_memory_table_" + get_uuid_str ()
3472- > >> >> >> v25 .8.12 .129 - lts
34733461
34743462 def execute_spark_query (query : str ):
34753463 return execute_spark_query_general (
@@ -3483,7 +3471,6 @@ def execute_spark_query(query: str):
34833471 execute_spark_query (
34843472 f"""
34853473 CREATE TABLE { TABLE_NAME } (
3486- <<<<<<< HEAD
34873474 tag INT,
34883475 date DATE,
34893476 date2 DATE,
@@ -3714,20 +3701,10 @@ def execute_spark_query(query: str, table_name):
37143701 USING iceberg
37153702 OPTIONS('format-version'='2')
37163703 """ , TABLE_NAME_2
3717- == == == =
3718- id INT ,
3719- data STRING
3720- )
3721- USING iceberg
3722- PARTITIONED BY (identity (id ))
3723- OPTIONS ('format-version' = '2' )
3724- """
3725- >>>>>>> v25.8.12.129-lts
37263704 )
37273705
37283706 execute_spark_query (
37293707 f"""
3730- << << << < HEAD
37313708 INSERT INTO { TABLE_NAME_2 } VALUES
37323709 (1, 'dow'),
37333710 (2, 'sparrow')
@@ -3871,7 +3848,41 @@ def test_system_tables_partition_sorting_keys(started_cluster, storage_type):
38713848 """ ).strip ().lower ()
38723849
38733850 assert res == '"bucket(16, id), day(ts)","id desc, hour(ts) asc"'
3874- == == == =
3851+
3852+
3853+ @pytest .mark .parametrize (
3854+ "storage_type" ,
3855+ ["s3" , "azure" , "local" ],
3856+ )
3857+ def test_partition_pruning_with_subquery_set (started_cluster , storage_type ):
3858+ instance = started_cluster .instances ["node1" ]
3859+ spark = started_cluster .spark_session
3860+ TABLE_NAME = "test_partition_pruning_" + storage_type + "_" + get_uuid_str ()
3861+ IN_MEMORY_TABLE = "in_memory_table_" + get_uuid_str ()
3862+
3863+ def execute_spark_query (query : str ):
3864+ return execute_spark_query_general (
3865+ spark ,
3866+ started_cluster ,
3867+ storage_type ,
3868+ TABLE_NAME ,
3869+ query ,
3870+ )
3871+
3872+ execute_spark_query (
3873+ f"""
3874+ CREATE TABLE { TABLE_NAME } (
3875+ id INT,
3876+ data STRING
3877+ )
3878+ USING iceberg
3879+ PARTITIONED BY (identity(id))
3880+ OPTIONS('format-version'='2')
3881+ """
3882+ )
3883+
3884+ execute_spark_query (
3885+ f"""
38753886 INSERT INTO { TABLE_NAME } VALUES
38763887 (1, 'a'),
38773888 (2, 'b'),
@@ -3881,15 +3892,13 @@ def test_system_tables_partition_sorting_keys(started_cluster, storage_type):
38813892 """
38823893 )
38833894
3884-
38853895 creation_expression = get_creation_expression (
38863896 storage_type , TABLE_NAME , started_cluster , table_function = True
38873897 )
38883898
38893899 instance .query (f"CREATE TABLE { IN_MEMORY_TABLE } (id INT) ENGINE = Memory" )
38903900 instance .query (f"INSERT INTO { IN_MEMORY_TABLE } VALUES (2), (4)" )
38913901
3892-
38933902 def check_validity_and_get_prunned_files (select_expression ):
38943903 settings1 = {
38953904 "use_iceberg_partition_pruning" : 0
@@ -3907,4 +3916,3 @@ def check_validity_and_get_prunned_files(select_expression):
39073916 )
39083917 == 3
39093918 )
3910- >>>>>>> v25.8.12.129-lts
0 commit comments