Skip to content

DDP Documentation Clarification #20647

@Unturned3

Description

@Unturned3

📚 Documentation

The documentation on DDP currently says:

Using DDP this way has a few disadvantages over torch.multiprocessing.spawn():

  1. All processes (including the main process) participate in training and have the updated state of the model and Trainer state.
  2. No multiprocessing pickle errors
  3. Easily scales to multi-node training

Are these meant to be advantages instead of disadvantages?

cc @lantiga @Borda

Metadata

Metadata

Assignees

No one assigned

    Labels

    docsDocumentation relatedneeds triageWaiting to be triaged by maintainers

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions