2525 FileUploadCompletionBody ,
2626 FileUploadData ,
2727 UploadedPart ,
28- ClientFileToProgramJob ,
28+ UserFileToProgramJob ,
2929 UserFile ,
3030)
3131from urllib .parse import urljoin
@@ -164,8 +164,8 @@ async def upload_file_to_program_job_async(
164164 # if a file has the same sha256 checksum
165165 # and name they are considered equal
166166 return file_result
167- user_file = UserFile (
168- ClientFileToProgramJob (
167+ user_file = ClientFile (
168+ UserFileToProgramJob (
169169 filename = file .name ,
170170 filesize = file .stat ().st_size ,
171171 sha256_checksum = checksum ,
@@ -214,8 +214,8 @@ async def upload_file_async(
214214 # if a file has the same sha256 checksum
215215 # and name they are considered equal
216216 return file_result
217- user_file = UserFile (
218- ClientFile (
217+ user_file = ClientFile (
218+ UserFile (
219219 filename = file .name ,
220220 filesize = file .stat ().st_size ,
221221 sha256_checksum = checksum ,
@@ -227,28 +227,30 @@ async def upload_file_async(
227227
228228 async def _upload_user_file (
229229 self ,
230- user_file : UserFile ,
230+ client_file : ClientFile ,
231231 file : Path ,
232232 timeout_seconds : int = DEFAULT_TIMEOUT_SECONDS ,
233233 max_concurrent_uploads : int = _MAX_CONCURRENT_UPLOADS ,
234234 ** kwargs ,
235235 ) -> File :
236236 assert file .is_file () # nosec
237237 client_upload_schema : ClientFileUploadData = super ().get_upload_links (
238- user_file = user_file , _request_timeout = timeout_seconds , ** kwargs
238+ client_file = client_file , _request_timeout = timeout_seconds , ** kwargs
239239 )
240240 chunk_size : int = client_upload_schema .upload_schema .chunk_size
241241 links : FileUploadData = client_upload_schema .upload_schema .links
242242 url_iter : Iterator [Tuple [int , str ]] = enumerate (
243243 iter (client_upload_schema .upload_schema .urls ), start = 1
244244 )
245245 n_urls : int = len (client_upload_schema .upload_schema .urls )
246- if n_urls < math .ceil (user_file .actual_instance .filesize / chunk_size ):
246+ if n_urls < math .ceil (client_file .actual_instance .filesize / chunk_size ):
247247 raise RuntimeError (
248248 "Did not receive sufficient number of upload URLs from the server."
249249 )
250250
251- abort_body = BodyAbortMultipartUploadV0FilesFileIdAbortPost (user_file = user_file )
251+ abort_body = BodyAbortMultipartUploadV0FilesFileIdAbortPost (
252+ client_file = client_file
253+ )
252254 upload_tasks : Set [asyncio .Task ] = set ()
253255 uploaded_parts : List [UploadedPart ] = []
254256
@@ -301,7 +303,7 @@ async def _upload_user_file(
301303 server_file : File = await self ._complete_multipart_upload (
302304 api_server_session ,
303305 links .complete_upload , # type: ignore
304- user_file ,
306+ client_file ,
305307 uploaded_parts ,
306308 )
307309 _logger .debug ("File upload complete: %s" , file .name )
@@ -311,11 +313,11 @@ async def _complete_multipart_upload(
311313 self ,
312314 http_client : AsyncHttpClient ,
313315 complete_link : str ,
314- user_file : UserFile ,
316+ client_file : ClientFile ,
315317 uploaded_parts : List [UploadedPart ],
316318 ) -> File :
317319 complete_payload = BodyCompleteMultipartUploadV0FilesFileIdCompletePost (
318- user_file = user_file ,
320+ client_file = client_file ,
319321 uploaded_parts = FileUploadCompletionBody (parts = uploaded_parts ),
320322 )
321323 response : Response = await http_client .post (
0 commit comments