File tree Expand file tree Collapse file tree 3 files changed +9
-9
lines changed Expand file tree Collapse file tree 3 files changed +9
-9
lines changed Original file line number Diff line number Diff line change @@ -77,14 +77,14 @@ def validate_source_dir(script, directory):
7777
7878
7979def get_mp_parameters (distribution ):
80- """Get the model parallelism parameters provided by the user
80+ """Get the model parallelism parameters provided by the user.
8181
8282 Args:
83- distribution: distribution dictionary defined by the user
83+ distribution: distribution dictionary defined by the user.
8484
8585 Returns:
8686 params: dictionary containing model parallelism parameters
87- to be used for training
87+ used for training.
8888 """
8989 try :
9090 mp_dict = distribution ["smdistributed" ]["modelparallel" ]
Original file line number Diff line number Diff line change @@ -103,9 +103,9 @@ def __init__(
103103 ``image_uri`` is required. If also ``None``, then a ``ValueError``
104104 will be raised.
105105 distribution (dict): A dictionary with information on how to run distributed training
106- (default: None). Currently we support distributed training with parameter servers,
107- Model Parallelism, Data Parallelism, and MPI. Model Parallelism can only be used
108- with MPI.
106+ (default: None). Currently, the following are supported:
107+ distributed training with parameter servers, SageMaker Distributed (SMD) Data
108+ and Model Parallelism, and MPI. SMD Model Parallelism can only be used with MPI.
109109 To enable parameter server use the following setup:
110110
111111 .. code:: python
Original file line number Diff line number Diff line change @@ -81,9 +81,9 @@ def __init__(
8181 ``image_uri`` is required. If also ``None``, then a ``ValueError``
8282 will be raised.
8383 distribution (dict): A dictionary with information on how to run distributed training
84- (default: None). Currently we support distributed training with parameter servers,
85- Model Parallelism, Data Parallelism, and MPI. Model Parallelism can only be used
86- with MPI.
84+ (default: None). Currently, the following are supported:
85+ distributed training with parameter servers, SageMaker Distributed (SMD) Data
86+ and Model Parallelism, and MPI. SMD Model Parallelism can only be used with MPI.
8787 To enable parameter server use the following setup:
8888
8989 .. code:: python
You can’t perform that action at this time.
0 commit comments