@@ -682,7 +682,7 @@ def to_csv(self,
682682 """
683683 if serde not in Pandas .VALID_CSV_SERDES :
684684 raise InvalidSerDe (f"{ serde } in not in the valid SerDe list ({ Pandas .VALID_CSV_SERDES } )" )
685- extra_args = {"sep" : sep , "serde" : serde , "escapechar" : escapechar }
685+ extra_args : Dict [ str , Optional [ str ]] = {"sep" : sep , "serde" : serde , "escapechar" : escapechar }
686686 return self .to_s3 (dataframe = dataframe ,
687687 path = path ,
688688 file_format = "csv" ,
@@ -767,7 +767,7 @@ def to_s3(self,
767767 procs_cpu_bound = None ,
768768 procs_io_bound = None ,
769769 cast_columns = None ,
770- extra_args = None ,
770+ extra_args : Optional [ Dict [ str , Optional [ str ]]] = None ,
771771 inplace : bool = True ,
772772 description : Optional [str ] = None ,
773773 parameters : Optional [Dict [str , str ]] = None ,
@@ -922,7 +922,7 @@ def _data_to_s3_dataset_writer(dataframe: pd.DataFrame,
922922 session_primitives : "SessionPrimitives" ,
923923 file_format : str ,
924924 cast_columns = None ,
925- extra_args = None ,
925+ extra_args : Optional [ Dict [ str , Optional [ str ]]] = None ,
926926 isolated_dataframe : bool = False ):
927927 objects_paths = []
928928 dataframe = Pandas ._cast_pandas (dataframe = dataframe , cast_columns = cast_columns )
@@ -980,7 +980,7 @@ def _data_to_s3_dataset_writer_remote(send_pipe,
980980 session_primitives : "SessionPrimitives" ,
981981 file_format ,
982982 cast_columns = None ,
983- extra_args = None ):
983+ extra_args : Optional [ Dict [ str , Optional [ str ]]] = None ):
984984 send_pipe .send (
985985 Pandas ._data_to_s3_dataset_writer (dataframe = dataframe ,
986986 path = path ,
@@ -996,35 +996,35 @@ def _data_to_s3_dataset_writer_remote(send_pipe,
996996
997997 @staticmethod
998998 def _data_to_s3_object_writer (dataframe : pd .DataFrame ,
999- path : " str" ,
999+ path : str ,
10001000 preserve_index : bool ,
1001- compression ,
1001+ compression : str ,
10021002 session_primitives : "SessionPrimitives" ,
1003- file_format ,
1004- cast_columns = None ,
1005- extra_args = None ,
1006- isolated_dataframe = False ):
1003+ file_format : str ,
1004+ cast_columns : Optional [ List [ str ]] = None ,
1005+ extra_args : Optional [ Dict [ str , Optional [ str ]]] = None ,
1006+ isolated_dataframe = False ) -> str :
10071007 fs = get_fs (session_primitives = session_primitives )
10081008 fs = pa .filesystem ._ensure_filesystem (fs )
10091009 mkdir_if_not_exists (fs , path )
10101010
10111011 if compression is None :
1012- compression_end = ""
1012+ compression_extension : str = ""
10131013 elif compression == "snappy" :
1014- compression_end = ".snappy"
1014+ compression_extension = ".snappy"
10151015 elif compression == "gzip" :
1016- compression_end = ".gz"
1016+ compression_extension = ".gz"
10171017 else :
10181018 raise InvalidCompression (compression )
10191019
1020- guid = pa .compat .guid ()
1020+ guid : str = pa .compat .guid ()
10211021 if file_format == "parquet" :
1022- outfile = f"{ guid } .parquet { compression_end } "
1022+ outfile : str = f"{ guid } { compression_extension } .parquet "
10231023 elif file_format == "csv" :
1024- outfile = f"{ guid } .csv { compression_end } "
1024+ outfile = f"{ guid } { compression_extension } .csv "
10251025 else :
10261026 raise UnsupportedFileFormat (file_format )
1027- object_path = "/" .join ([path , outfile ])
1027+ object_path : str = "/" .join ([path , outfile ])
10281028 if file_format == "parquet" :
10291029 Pandas .write_parquet_dataframe (dataframe = dataframe ,
10301030 path = object_path ,
0 commit comments