Skip to content

Should the total epoch size be less when using multi-gpu DDP? #7175

Discussion options

You must be logged in to vote

Hi @jipson7 ,
First of all: You're right, that's how it should be.

We tried to reproduce this, but for us this produced the following (correct) output. Do you have a minimal reproduction example?

Epoch 0: 100%|███████████████████████████████████████████████████████| 10/10 [00:00<00:00, 17.23it/s, loss=-43.6, v_num=272]
seen train: 5
seen train: 5

Replies: 2 comments 6 replies

Comment options

You must be logged in to vote
4 replies
@jipson7
Comment options

@jipson7
Comment options

@jipson7
Comment options

@awaelchli
Comment options

Answer selected by jipson7
Comment options

You must be logged in to vote
2 replies
@justusschock
Comment options

@jipson7
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
3 participants