Replies: 1 comment 1 reply
-
Great question! We actually do something like this for the PovertyMap dataset, so perhaps that would be a helpful reference? https://github.com/p-lambda/wilds/blob/main/wilds/datasets/poverty_dataset.py |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hello and thank you for this amazing package.
Instead of using replicates, I would be interested in adding a cross validation training and evaluation scheme based on the domain metadata.
Say a dataset has domain: A,B,C. I would like to:
Finally average the in distribution and the out of distribution metric to have the final performance.
Here the 70-30 split is arbitrary and should be modifiable.
I am just starting exploring the package having only replicated the ERM result on the camelyon17 dataset.
It seems that the grouper object might be a good start to implement the following procedure. But, I am still lacking a high level overview of the code. So how would you do this ?
Beta Was this translation helpful? Give feedback.
All reactions