@@ -2535,7 +2535,7 @@ def spark() -> "SparkSession":
25352535 spark_version = "." .join (importlib .metadata .version ("pyspark" ).split ("." )[:2 ])
25362536 scala_version = "2.13"
25372537 iceberg_version = "1.10.0"
2538- hadoop_version = "3.3.4 "
2538+ hadoop_version = "3.3.6 "
25392539 aws_sdk_version = "1.12.753"
25402540
25412541 os .environ ["PYSPARK_SUBMIT_ARGS" ] = (
@@ -2577,6 +2577,11 @@ def spark() -> "SparkSession":
25772577 .config ("spark.sql.catalog.spark_catalog.warehouse" , "s3://warehouse/hive/" )
25782578 .config ("spark.hadoop.fs.s3a.endpoint" , "http://localhost:9000" )
25792579 .config ("spark.hadoop.fs.s3a.path.style.access" , "true" )
2580+ .config ("spark.hadoop.fs.s3a.threads.keepalivetime" , "60000" )
2581+ .config ("spark.hadoop.fs.s3a.connection.establish.timeout" , "30000" )
2582+ .config ("spark.hadoop.fs.s3a.connection.timeout" , "200000" )
2583+ .config ("spark.hadoop.fs.s3a.multipart.purge.age" , str (24 * 60 * 60 ))
2584+ .config ("spark.hadoop.fs.s3a.aws.credentials.provider" , "com.amazonaws.auth.EnvironmentVariableCredentialsProvider" )
25802585 .config ("spark.sql.catalogImplementation" , "hive" )
25812586 .config ("spark.sql.defaultCatalog" , "integration" )
25822587 .config ("spark.sql.execution.arrow.pyspark.enabled" , "true" )
0 commit comments