@@ -343,7 +343,7 @@ def process_message(self, connector, host, secret_key, resource, parameters):
343343 try :
344344 # upload metadata to the processed file or dataset
345345 if 'metadata' in result .keys ():
346- self .logger .info ("upload metadata" )
346+ self .logger .debug ("upload metadata" )
347347 if type == 'file' :
348348 metadata = self .get_metadata (result .get ('metadata' ), 'file' , file_id , host )
349349 self .logger .debug (metadata )
@@ -357,19 +357,18 @@ def process_message(self, connector, host, secret_key, resource, parameters):
357357
358358 # upload previews to the processed file
359359 if 'previews' in result .keys ():
360- self .logger .info ("upload previews" )
361360 if type == 'file' :
362361 for preview in result ['previews' ]:
363362 if os .path .exists (str (preview )):
364363 preview = {'file' : preview }
365- self .logger .info ("upload preview" )
364+ self .logger .debug ("upload preview" )
366365 pyclowder .files .upload_preview (connector , host , secret_key , file_id , str (preview ))
367366 else :
368367 # TODO: Add Clowder endpoint (& pyclowder method) to attach previews to datasets
369368 self .logger .error ("previews not currently supported for resource type: %s" % type )
370369
371370 if 'tags' in result .keys ():
372- self .logger .info ("upload tags" )
371+ self .logger .debug ("upload tags" )
373372 tags = {"tags" : result ["tags" ]}
374373 if type == 'file' :
375374 pyclowder .files .upload_tags (connector , host , secret_key , file_id , tags )
@@ -378,7 +377,7 @@ def process_message(self, connector, host, secret_key, resource, parameters):
378377
379378 # upload output files to the processed file's parent dataset or processed dataset
380379 if 'outputs' in result .keys ():
381- self .logger .info ("upload output files" )
380+ self .logger .debug ("upload output files" )
382381 if type == 'file' or type == 'dataset' :
383382 for output in result ['outputs' ]:
384383 if os .path .exists (str (output )):
@@ -395,16 +394,16 @@ def process_message(self, connector, host, secret_key, resource, parameters):
395394 description = nds ['description' ] if 'description' in nds .keys () else ""
396395 new_dataset_id = pyclowder .datasets .create_empty (connector , host , secret_key , nds ['name' ],
397396 description )
398- self .logger .info ("created new dataset: %s" % new_dataset_id )
397+ self .logger .debug ("created new dataset: %s" % new_dataset_id )
399398
400399 if 'metadata' in nds .keys ():
401- self .logger .info ("upload metadata to new dataset" )
400+ self .logger .debug ("upload metadata to new dataset" )
402401 metadata = self .get_metadata (nds .get ('metadata' ), 'dataset' , new_dataset_id , host )
403402 self .logger .debug (metadata )
404403 pyclowder .datasets .upload_metadata (connector , host , secret_key , new_dataset_id , metadata )
405404
406405 if 'outputs' in nds .keys ():
407- self .logger .info ("upload output files to new dataset" )
406+ self .logger .debug ("upload output files to new dataset" )
408407 for output in nds ['outputs' ]:
409408 if os .path .exists (str (output )):
410409 pyclowder .files .upload_to_dataset (connector , host , secret_key , new_dataset_id ,
0 commit comments