@@ -68,14 +68,14 @@ def read_parquet(
6868 if table_partition_cols is None :
6969 table_partition_cols = []
7070 return SessionContext .global_ctx ().read_parquet (
71- str (path ),
72- table_partition_cols ,
73- parquet_pruning ,
74- file_extension ,
75- skip_metadata ,
76- schema ,
77- file_sort_order ,
78- )
71+ str (path ),
72+ table_partition_cols ,
73+ parquet_pruning ,
74+ file_extension ,
75+ skip_metadata ,
76+ schema ,
77+ file_sort_order ,
78+ )
7979
8080
8181def read_json (
@@ -108,13 +108,13 @@ def read_json(
108108 if table_partition_cols is None :
109109 table_partition_cols = []
110110 return SessionContext .global_ctx ().read_json (
111- str (path ),
112- schema ,
113- schema_infer_max_records ,
114- file_extension ,
115- table_partition_cols ,
116- file_compression_type ,
117- )
111+ str (path ),
112+ schema ,
113+ schema_infer_max_records ,
114+ file_extension ,
115+ table_partition_cols ,
116+ file_compression_type ,
117+ )
118118
119119
120120def read_csv (
@@ -157,15 +157,15 @@ def read_csv(
157157 path = [str (p ) for p in path ] if isinstance (path , list ) else str (path )
158158
159159 return SessionContext .global_ctx ().read_csv (
160- path ,
161- schema ,
162- has_header ,
163- delimiter ,
164- schema_infer_max_records ,
165- file_extension ,
166- table_partition_cols ,
167- file_compression_type ,
168- )
160+ path ,
161+ schema ,
162+ has_header ,
163+ delimiter ,
164+ schema_infer_max_records ,
165+ file_extension ,
166+ table_partition_cols ,
167+ file_compression_type ,
168+ )
169169
170170
171171def read_avro (
@@ -192,5 +192,5 @@ def read_avro(
192192 if file_partition_cols is None :
193193 file_partition_cols = []
194194 return SessionContext .global_ctx ().read_avro (
195- str (path ), schema , file_partition_cols , file_extension
196- )
195+ str (path ), schema , file_partition_cols , file_extension
196+ )
0 commit comments