Collating all files at the end of a pipeline #2925
-
Hi All, Interested in what people do in the following scenario. Say we have 10 processes in our DSL-1 main.nf pipeline. Currently what we are doing is, once the pipeline is complete, we run a python script that collates select data from all of the output files and puts it into an excel file for easier review by genomic analysts. We would like to incorporate that last step into the pipeline but it seems that there are variable practices in how to do this in NF. So, I'm curious what other people are doing for steps like this. Thanks! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 5 replies
-
It depends on which outputs you need. If you only need to collect outputs published by Nextflow processes, then you can just add another process to your workflow that collects those outputs. But if you also need the reports created by Nextflow (execution report, timeline, trace, log), then you must do that outside of Nextflow itself. The best thing you can do in the latter case is to wrap the two steps -- Nextflow pipeline and python script -- in a bash script or even a Nextflow meta-pipeline (I have done this myself). |
Beta Was this translation helpful? Give feedback.
It depends on which outputs you need. If you only need to collect outputs published by Nextflow processes, then you can just add another process to your workflow that collects those outputs. But if you also need the reports created by Nextflow (execution report, timeline, trace, log), then you must do that outside of Nextflow itself.
The best thing you can do in the latter case is to wrap the two steps -- Nextflow pipeline and python script -- in a bash script or even a Nextflow meta-pipeline (I have done this myself).