Skip to content

Commit 7d556a0

Browse files
committed
use environment variable
1 parent 70861e2 commit 7d556a0

File tree

3 files changed

+26
-25
lines changed

3 files changed

+26
-25
lines changed

pyclowder/datasets.py

Lines changed: 12 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@
1717

1818
clowder_version = int(os.getenv('CLOWDER_VERSION', '1'))
1919

20-
def create_empty(connector, host, key, datasetname, description, parentid=None, spaceid=None, clowder_version=1):
20+
def create_empty(connector, host, key, datasetname, description, parentid=None, spaceid=None):
2121
"""Create a new dataset in Clowder.
2222
2323
Keyword arguments:
@@ -37,7 +37,7 @@ def create_empty(connector, host, key, datasetname, description, parentid=None,
3737
return datasetid
3838

3939

40-
def delete(connector, host, key, datasetid, clowder_version=1):
40+
def delete(connector, host, key, datasetid):
4141
"""Delete dataset from Clowder.
4242
4343
Keyword arguments:
@@ -56,7 +56,7 @@ def delete(connector, host, key, datasetid, clowder_version=1):
5656
return json.loads(result.text)
5757

5858

59-
def delete_by_collection(connector, host, key, collectionid, recursive=True, delete_colls=False, clowder_version=1):
59+
def delete_by_collection(connector, host, key, collectionid, recursive=True, delete_colls=False):
6060
"""Delete datasets from Clowder by iterating through collection.
6161
6262
Keyword arguments:
@@ -81,7 +81,7 @@ def delete_by_collection(connector, host, key, collectionid, recursive=True, del
8181
delete_collection(connector, client, collectionid)
8282

8383

84-
def download(connector, host, key, datasetid, clowder_version=1):
84+
def download(connector, host, key, datasetid):
8585
"""Download dataset to be processed from Clowder as zip file.
8686
8787
Keyword arguments:
@@ -98,7 +98,7 @@ def download(connector, host, key, datasetid, clowder_version=1):
9898
return zipfile
9999

100100

101-
def download_metadata(connector, host, key, datasetid, extractor=None, clowder_version=1):
101+
def download_metadata(connector, host, key, datasetid, extractor=None):
102102
"""Download dataset JSON-LD metadata from Clowder.
103103
104104
Keyword arguments:
@@ -117,7 +117,7 @@ def download_metadata(connector, host, key, datasetid, extractor=None, clowder_v
117117
return result_json
118118

119119

120-
def get_info(connector, host, key, datasetid, clowder_version=1):
120+
def get_info(connector, host, key, datasetid):
121121
"""Get basic dataset information from UUID.
122122
123123
Keyword arguments:
@@ -134,7 +134,7 @@ def get_info(connector, host, key, datasetid, clowder_version=1):
134134
return info
135135

136136

137-
def get_file_list(connector, client, datasetid, clowder_version=1):
137+
def get_file_list(connector, client, datasetid):
138138
"""Get list of files in a dataset as JSON object.
139139
140140
Keyword arguments:
@@ -150,7 +150,7 @@ def get_file_list(connector, client, datasetid, clowder_version=1):
150150
return file_list
151151

152152

153-
def remove_metadata(connector, host, key, datasetid, extractor=None, clowder_version=1):
153+
def remove_metadata(connector, host, key, datasetid, extractor=None):
154154
"""Delete dataset JSON-LD metadata from Clowder.
155155
156156
Keyword arguments:
@@ -168,7 +168,7 @@ def remove_metadata(connector, host, key, datasetid, extractor=None, clowder_ver
168168
v1datasets.remove_metadata(connector, client, datasetid, extractor)
169169

170170

171-
def submit_extraction(connector, host, key, datasetid, extractorname, clowder_version=1):
171+
def submit_extraction(connector, host, key, datasetid, extractorname):
172172
"""Submit dataset for extraction by given extractor.
173173
174174
Keyword arguments:
@@ -185,7 +185,7 @@ def submit_extraction(connector, host, key, datasetid, extractorname, clowder_ve
185185
result_status_code = v1datasets.submit_extraction(connector, client, datasetid, extractorname)
186186

187187

188-
def submit_extractions_by_collection(connector, host, key, collectionid, extractorname, recursive=True, clowder_version=1):
188+
def submit_extractions_by_collection(connector, host, key, collectionid, extractorname, recursive=True):
189189
"""Manually trigger an extraction on all datasets in a collection.
190190
191191
This will iterate through all datasets in the given collection and submit them to
@@ -211,7 +211,7 @@ def submit_extractions_by_collection(connector, host, key, collectionid, extract
211211
submit_extractions_by_collection(connector, client, coll['id'], extractorname, recursive)
212212

213213

214-
def upload_tags(connector, host, key, datasetid, tags, clowder_version=1):
214+
def upload_tags(connector, host, key, datasetid, tags):
215215
"""Upload dataset tag to Clowder.
216216
217217
Keyword arguments:
@@ -230,7 +230,7 @@ def upload_tags(connector, host, key, datasetid, tags, clowder_version=1):
230230
verify=connector.ssl_verify if connector else True)
231231

232232

233-
def upload_metadata(connector, host, key, datasetid, metadata, clowder_version=1):
233+
def upload_metadata(connector, host, key, datasetid, metadata):
234234
"""Upload dataset JSON-LD metadata to Clowder.
235235
236236
Keyword arguments:

pyclowder/extractors.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,7 @@
2424
import pyclowder.files
2525
import pyclowder.datasets
2626

27+
clowder_version = int(os.getenv('CLOWDER_VERSION', '1'))
2728

2829
class Extractor(object):
2930
"""Basic extractor.
@@ -228,7 +229,7 @@ def _get_extractor_info_v2(self):
228229
return current_extractor_info
229230

230231

231-
def get_metadata(self, content, resource_type, resource_id, server=None, clowder_version=1):
232+
def get_metadata(self, content, resource_type, resource_id, server=None):
232233
"""Generate a metadata field.
233234
234235
This will return a metadata dict that is valid JSON-LD. This will use the results as well as the information

pyclowder/files.py

Lines changed: 12 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@
2929

3030

3131
# pylint: disable=too-many-arguments
32-
def download(connector, host, key, fileid, intermediatefileid=None, ext="", clowder_version=1):
32+
def download(connector, host, key, fileid, intermediatefileid=None, ext=""):
3333
"""Download file to be processed from Clowder.
3434
3535
Keyword arguments:
@@ -48,7 +48,7 @@ def download(connector, host, key, fileid, intermediatefileid=None, ext="", clow
4848
return inputfilename
4949

5050

51-
def download_info(connector, host, key, fileid, clowder_version=1):
51+
def download_info(connector, host, key, fileid):
5252
"""Download file summary metadata from Clowder.
5353
5454
Keyword arguments:
@@ -65,7 +65,7 @@ def download_info(connector, host, key, fileid, clowder_version=1):
6565
return result.json()
6666

6767

68-
def download_metadata(connector, host, key, fileid, extractor=None, clowder_version=1):
68+
def download_metadata(connector, host, key, fileid, extractor=None):
6969
"""Download file JSON-LD metadata from Clowder.
7070
7171
Keyword arguments:
@@ -83,7 +83,7 @@ def download_metadata(connector, host, key, fileid, extractor=None, clowder_vers
8383
return result.json()
8484

8585

86-
def submit_extraction(connector, host, key, fileid, extractorname, clowder_version=1):
86+
def submit_extraction(connector, host, key, fileid, extractorname):
8787
"""Submit file for extraction by given extractor.
8888
8989
Keyword arguments:
@@ -101,7 +101,7 @@ def submit_extraction(connector, host, key, fileid, extractorname, clowder_versi
101101
return result.json()
102102

103103

104-
def submit_extractions_by_dataset(connector, host, key, datasetid, extractorname, ext=False, clowder_version=1):
104+
def submit_extractions_by_dataset(connector, host, key, datasetid, extractorname, ext=False):
105105
"""Manually trigger an extraction on all files in a dataset.
106106
107107
This will iterate through all files in the given dataset and submit them to
@@ -126,7 +126,7 @@ def submit_extractions_by_dataset(connector, host, key, datasetid, extractorname
126126
submit_extraction(connector, host, key, f['id'], extractorname)
127127

128128

129-
def submit_extractions_by_collection(connector, host, key, collectionid, extractorname, ext=False, recursive=True, clowder_version=1):
129+
def submit_extractions_by_collection(connector, host, key, collectionid, extractorname, ext=False, recursive=True):
130130
"""Manually trigger an extraction on all files in a collection.
131131
132132
This will iterate through all datasets in the given collection and submit them to
@@ -153,7 +153,7 @@ def submit_extractions_by_collection(connector, host, key, collectionid, extract
153153
submit_extractions_by_collection(connector, host, key, coll['id'], extractorname, ext, recursive)
154154

155155

156-
def upload_metadata(connector, host, key, fileid, metadata, clowder_version=1):
156+
def upload_metadata(connector, host, key, fileid, metadata):
157157
"""Upload file JSON-LD metadata to Clowder.
158158
159159
Keyword arguments:
@@ -171,7 +171,7 @@ def upload_metadata(connector, host, key, fileid, metadata, clowder_version=1):
171171

172172

173173
# pylint: disable=too-many-arguments
174-
def upload_preview(connector, host, key, fileid, previewfile, previewmetadata=None, preview_mimetype=None, clowder_version=1):
174+
def upload_preview(connector, host, key, fileid, previewfile, previewmetadata=None, preview_mimetype=None):
175175
"""Upload preview to Clowder.
176176
177177
Keyword arguments:
@@ -219,7 +219,7 @@ def upload_preview(connector, host, key, fileid, previewfile, previewmetadata=No
219219
return previewid
220220

221221

222-
def upload_tags(connector, host, key, fileid, tags, clowder_version=1):
222+
def upload_tags(connector, host, key, fileid, tags):
223223
"""Upload file tag to Clowder.
224224
225225
Keyword arguments:
@@ -238,7 +238,7 @@ def upload_tags(connector, host, key, fileid, tags, clowder_version=1):
238238
verify=connector.ssl_verify if connector else True)
239239

240240

241-
def upload_thumbnail(connector, host, key, fileid, thumbnail, clowder_version=1):
241+
def upload_thumbnail(connector, host, key, fileid, thumbnail):
242242
"""Upload thumbnail to Clowder.
243243
244244
Keyword arguments:
@@ -267,7 +267,7 @@ def upload_thumbnail(connector, host, key, fileid, thumbnail, clowder_version=1)
267267
return thumbnailid
268268

269269

270-
def upload_to_dataset(connector, host, key, datasetid, filepath, check_duplicate=False, clowder_version=1):
270+
def upload_to_dataset(connector, host, key, datasetid, filepath, check_duplicate=False):
271271
"""Upload file to existing Clowder dataset.
272272
273273
Keyword arguments:
@@ -313,7 +313,7 @@ def upload_to_dataset(connector, host, key, datasetid, filepath, check_duplicate
313313
logger.error("unable to upload file %s (not found)", filepath)
314314

315315

316-
def _upload_to_dataset_local(connector, host, key, datasetid, filepath, clowder_version=1):
316+
def _upload_to_dataset_local(connector, host, key, datasetid, filepath):
317317
"""Upload file POINTER to existing Clowder dataset. Does not copy actual file bytes.
318318
319319
Keyword arguments:

0 commit comments

Comments
 (0)