Why does the self-supervised pre-training tutorial create two cropped views of the input image? #6659
Unanswered
ArjunNarayanan
asked this question in
Q&A
Replies: 1 comment 1 reply
-
This arg can be different, if GPU mem allows, it can be set to larger num samples or 1, but remember the contrastive learning might be impacted by the num sample and batch size. Thank you. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
I'm taking a look at the self-supervised pre-training tutorial. I was wondering about the data-augmentation in the transform pipeline. Specifically, the tutorial samples two cropped views of the input image,
I was wondering what is the purpose of this? It seems like this effectively doubles the batch-size where every subsequent image in the batch is drawn from the same input image. Thanks for any insights you can share on this!
Beta Was this translation helpful? Give feedback.
All reactions