Using customized ddp sampler get stuck during training #15249
Unanswered
GongXinyuu
asked this question in
DDP / multi-GPU / multi-node
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I have set
trainer. replace_sampler_ddp=False
and implement a very naiveDistributedSampler
as below:The training gets stuck without any error message. Before I do that, everything runs well.
Beta Was this translation helpful? Give feedback.
All reactions