2929
3030
3131# pylint: disable=too-many-arguments
32- def download (connector , host , key , fileid , intermediatefileid = None , ext = "" , clowder_version = 1 ):
32+ def download (connector , host , key , fileid , intermediatefileid = None , ext = "" ):
3333 """Download file to be processed from Clowder.
3434
3535 Keyword arguments:
@@ -48,7 +48,7 @@ def download(connector, host, key, fileid, intermediatefileid=None, ext="", clow
4848 return inputfilename
4949
5050
51- def download_info (connector , host , key , fileid , clowder_version = 1 ):
51+ def download_info (connector , host , key , fileid ):
5252 """Download file summary metadata from Clowder.
5353
5454 Keyword arguments:
@@ -65,7 +65,7 @@ def download_info(connector, host, key, fileid, clowder_version=1):
6565 return result .json ()
6666
6767
68- def download_metadata (connector , host , key , fileid , extractor = None , clowder_version = 1 ):
68+ def download_metadata (connector , host , key , fileid , extractor = None ):
6969 """Download file JSON-LD metadata from Clowder.
7070
7171 Keyword arguments:
@@ -83,7 +83,7 @@ def download_metadata(connector, host, key, fileid, extractor=None, clowder_vers
8383 return result .json ()
8484
8585
86- def submit_extraction (connector , host , key , fileid , extractorname , clowder_version = 1 ):
86+ def submit_extraction (connector , host , key , fileid , extractorname ):
8787 """Submit file for extraction by given extractor.
8888
8989 Keyword arguments:
@@ -101,7 +101,7 @@ def submit_extraction(connector, host, key, fileid, extractorname, clowder_versi
101101 return result .json ()
102102
103103
104- def submit_extractions_by_dataset (connector , host , key , datasetid , extractorname , ext = False , clowder_version = 1 ):
104+ def submit_extractions_by_dataset (connector , host , key , datasetid , extractorname , ext = False ):
105105 """Manually trigger an extraction on all files in a dataset.
106106
107107 This will iterate through all files in the given dataset and submit them to
@@ -126,7 +126,7 @@ def submit_extractions_by_dataset(connector, host, key, datasetid, extractorname
126126 submit_extraction (connector , host , key , f ['id' ], extractorname )
127127
128128
129- def submit_extractions_by_collection (connector , host , key , collectionid , extractorname , ext = False , recursive = True , clowder_version = 1 ):
129+ def submit_extractions_by_collection (connector , host , key , collectionid , extractorname , ext = False , recursive = True ):
130130 """Manually trigger an extraction on all files in a collection.
131131
132132 This will iterate through all datasets in the given collection and submit them to
@@ -153,7 +153,7 @@ def submit_extractions_by_collection(connector, host, key, collectionid, extract
153153 submit_extractions_by_collection (connector , host , key , coll ['id' ], extractorname , ext , recursive )
154154
155155
156- def upload_metadata (connector , host , key , fileid , metadata , clowder_version = 1 ):
156+ def upload_metadata (connector , host , key , fileid , metadata ):
157157 """Upload file JSON-LD metadata to Clowder.
158158
159159 Keyword arguments:
@@ -171,7 +171,7 @@ def upload_metadata(connector, host, key, fileid, metadata, clowder_version=1):
171171
172172
173173# pylint: disable=too-many-arguments
174- def upload_preview (connector , host , key , fileid , previewfile , previewmetadata = None , preview_mimetype = None , clowder_version = 1 ):
174+ def upload_preview (connector , host , key , fileid , previewfile , previewmetadata = None , preview_mimetype = None ):
175175 """Upload preview to Clowder.
176176
177177 Keyword arguments:
@@ -219,7 +219,7 @@ def upload_preview(connector, host, key, fileid, previewfile, previewmetadata=No
219219 return previewid
220220
221221
222- def upload_tags (connector , host , key , fileid , tags , clowder_version = 1 ):
222+ def upload_tags (connector , host , key , fileid , tags ):
223223 """Upload file tag to Clowder.
224224
225225 Keyword arguments:
@@ -238,7 +238,7 @@ def upload_tags(connector, host, key, fileid, tags, clowder_version=1):
238238 verify = connector .ssl_verify if connector else True )
239239
240240
241- def upload_thumbnail (connector , host , key , fileid , thumbnail , clowder_version = 1 ):
241+ def upload_thumbnail (connector , host , key , fileid , thumbnail ):
242242 """Upload thumbnail to Clowder.
243243
244244 Keyword arguments:
@@ -267,7 +267,7 @@ def upload_thumbnail(connector, host, key, fileid, thumbnail, clowder_version=1)
267267 return thumbnailid
268268
269269
270- def upload_to_dataset (connector , host , key , datasetid , filepath , check_duplicate = False , clowder_version = 1 ):
270+ def upload_to_dataset (connector , host , key , datasetid , filepath , check_duplicate = False ):
271271 """Upload file to existing Clowder dataset.
272272
273273 Keyword arguments:
@@ -313,7 +313,7 @@ def upload_to_dataset(connector, host, key, datasetid, filepath, check_duplicate
313313 logger .error ("unable to upload file %s (not found)" , filepath )
314314
315315
316- def _upload_to_dataset_local (connector , host , key , datasetid , filepath , clowder_version = 1 ):
316+ def _upload_to_dataset_local (connector , host , key , datasetid , filepath ):
317317 """Upload file POINTER to existing Clowder dataset. Does not copy actual file bytes.
318318
319319 Keyword arguments:
0 commit comments