Skip to content
Discussion options

You must be logged in to vote

The "querying n local files" step happens in targets:::runtiem_set_file_info(). The slow step is a (wrapped) file.info() call:

info <- file_info(

For most large pipelines, querying all the files together can actually speed up the pipeline. If you query them one by one, you'll get a lot more operating system overhead.

If that step is slow, you might try tar_prune() to remove unnecessary files from _targets/objects/. The only caveats with that are (1) if a collaborator is running a pipeline on the same data store as you, or (2) if you are using content-addressable storage (tar_repository_cas() e.g.).

Slowness might also come fr…

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@dblodgett-usgs
Comment options

Answer selected by dblodgett-usgs
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Help
Labels
None yet
2 participants