@@ -22,7 +22,7 @@ def get_worker_spark_session(
2222 """Construct Spark Session using run parameters and application settings"""
2323 from pyspark .sql import SparkSession
2424
25- name = run .transfer .group .name + "_" + run .transfer .name
25+ name = run .transfer .group .name + "_" + run .transfer .name # noqa: WPS336
2626 spark_builder = SparkSession .builder .appName (f"syncmaster_{ name } " )
2727
2828 for k , v in get_spark_session_conf (source , target ).items ():
@@ -35,15 +35,16 @@ def get_worker_spark_session(
3535 return spark_builder .getOrCreate ()
3636
3737
38- def get_packages (connection_type : str ) -> list [str ]:
38+ def get_packages (connection_type : str ) -> list [str ]: # noqa: WPS212
3939 import pyspark
4040 from onetl .connection import MSSQL , Clickhouse , MySQL , Oracle , Postgres , SparkS3
4141 from onetl .file .format import XML , Excel
4242
4343 # excel version is hardcoded due to https://github.com/nightscape/spark-excel/issues/902
44- file_formats_spark_packages : list [str ] = XML .get_packages (spark_version = pyspark .__version__ ) + Excel .get_packages (
45- spark_version = "3.5.1" ,
46- )
44+ file_formats_spark_packages : list [str ] = [
45+ * XML .get_packages (spark_version = pyspark .__version__ ),
46+ * Excel .get_packages (spark_version = "3.5.1" ),
47+ ]
4748
4849 if connection_type == "postgres" :
4950 return Postgres .get_packages ()
@@ -71,15 +72,10 @@ def get_packages(connection_type: str) -> list[str]:
7172 return []
7273
7374
74- def get_excluded_packages (db_type : str ):
75- if db_type == "s3" :
76- return [
77- "com.google.cloud.bigdataoss:gcs-connector" ,
78- "org.apache.hadoop:hadoop-aliyun" ,
79- "org.apache.hadoop:hadoop-azure-datalake" ,
80- "org.apache.hadoop:hadoop-azure" ,
81- ]
82- return []
75+ def get_excluded_packages (db_type : str ) -> list [str ]:
76+ from onetl .connection import SparkS3
77+
78+ return SparkS3 .get_exclude_packages ()
8379
8480
8581def get_spark_session_conf (
0 commit comments