Replies: 1 comment
-
|
Hi. This is an interesting use case. It should be doable, but I'd need to dig into the details, which will at least two weeks before I have the time. A workaround would be to set everything up as you'd be rerunning everything from scratch, but then detect if the results have already been computed and return early (e.g. with a It's not clear to me how you decide how something needs to be re-computed or not. It sounds like you know this information upfront. For instance, say your list of input data.frames is called Rather technical detouring comment: It would be nice if |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I'm using
future_mapfrom thefurrrpackage to parallelize a function I made to estimate a valueA(via bootstrapping) across a list of data frames (x), where each data frame corresponds to one organism. I set a single seed withfuture_options, however I ran into the problem that when I ran this same script on a subset of that list of data frames (meaning not all organisms where included) the estimatedAfor any organism was different than theAobtained for that same organism when I ran it on the full list.So I thought I could maybe generate a list of seeds, and have each seed assigned to each organism so that if I subset
xI can subset the seed list as well and always get the same results per organism regardless of the order/number of organisms in the data frame list. I tried this with the second method you showed for generating valid.Random.seedsequences, so I generated a list of valid seeds, whose length is the length of the unfilteredx. Each entry of the seed list gets assigned to an organism in a dataframe:So when I want to run my function on a subset of organisms I just filter the data frame above and get the filtered seeds list. But it didn't work, i.e. the
Avalue for Organism1 was different than theAvalue for Organism1 in the filteredxrun.Does anyone have any idea how to approach this problem? I hope I made myself clear, and also it could be that I'm just misunderstanding how the pre-generated seeds work. Would like to know what you think!
I guess I should also note that when I repeatedly run the same
x(filtered or non-filtered, but just keeping it the same) theAvalues stay the same, which I suppose is expected. Also another note is the reason why I'm subsetting is to reduce the number of hypothesis tests done inside my function, but that doesn't have anything to do with estimatingA.Beta Was this translation helpful? Give feedback.
All reactions