2222from pyclowder .connectors import RabbitMQConnector , HPCConnector , LocalConnector
2323from pyclowder .utils import CheckMessage , setup_logging
2424import pyclowder .files
25+ import pyclowder .datasets
2526
2627
2728class Extractor (object ):
@@ -342,7 +343,7 @@ def process_message(self, connector, host, secret_key, resource, parameters):
342343 try :
343344 # upload metadata to the processed file or dataset
344345 if 'metadata' in result .keys ():
345- self .logger .info ("upload metadata" )
346+ self .logger .debug ("upload metadata" )
346347 if type == 'file' :
347348 metadata = self .get_metadata (result .get ('metadata' ), 'file' , file_id , host )
348349 self .logger .debug (metadata )
@@ -356,20 +357,27 @@ def process_message(self, connector, host, secret_key, resource, parameters):
356357
357358 # upload previews to the processed file
358359 if 'previews' in result .keys ():
359- self .logger .info ("upload previews" )
360360 if type == 'file' :
361361 for preview in result ['previews' ]:
362362 if os .path .exists (str (preview )):
363363 preview = {'file' : preview }
364- self .logger .info ("upload preview" )
364+ self .logger .debug ("upload preview" )
365365 pyclowder .files .upload_preview (connector , host , secret_key , file_id , str (preview ))
366366 else :
367367 # TODO: Add Clowder endpoint (& pyclowder method) to attach previews to datasets
368368 self .logger .error ("previews not currently supported for resource type: %s" % type )
369369
370+ if 'tags' in result .keys ():
371+ self .logger .debug ("upload tags" )
372+ tags = {"tags" : result ["tags" ]}
373+ if type == 'file' :
374+ pyclowder .files .upload_tags (connector , host , secret_key , file_id , tags )
375+ else :
376+ pyclowder .datasets .upload_tags (connector , host , secret_key , dataset_id , tags )
377+
370378 # upload output files to the processed file's parent dataset or processed dataset
371379 if 'outputs' in result .keys ():
372- self .logger .info ("upload output files" )
380+ self .logger .debug ("upload output files" )
373381 if type == 'file' or type == 'dataset' :
374382 for output in result ['outputs' ]:
375383 if os .path .exists (str (output )):
@@ -386,16 +394,16 @@ def process_message(self, connector, host, secret_key, resource, parameters):
386394 description = nds ['description' ] if 'description' in nds .keys () else ""
387395 new_dataset_id = pyclowder .datasets .create_empty (connector , host , secret_key , nds ['name' ],
388396 description )
389- self .logger .info ("created new dataset: %s" % new_dataset_id )
397+ self .logger .debug ("created new dataset: %s" % new_dataset_id )
390398
391399 if 'metadata' in nds .keys ():
392- self .logger .info ("upload metadata to new dataset" )
400+ self .logger .debug ("upload metadata to new dataset" )
393401 metadata = self .get_metadata (nds .get ('metadata' ), 'dataset' , new_dataset_id , host )
394402 self .logger .debug (metadata )
395403 pyclowder .datasets .upload_metadata (connector , host , secret_key , new_dataset_id , metadata )
396404
397405 if 'outputs' in nds .keys ():
398- self .logger .info ("upload output files to new dataset" )
406+ self .logger .debug ("upload output files to new dataset" )
399407 for output in nds ['outputs' ]:
400408 if os .path .exists (str (output )):
401409 pyclowder .files .upload_to_dataset (connector , host , secret_key , new_dataset_id ,
0 commit comments