Skip to content

Commit a2322cc

Browse files
committed
Extra edits
1 parent d153df9 commit a2322cc

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

articles/machine-learning/concept-distributed-training.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -27,15 +27,15 @@ There are two main types of distributed training: [data parallelism](#data-paral
2727

2828
* [Distributed training with TensorFlow](how-to-train-distributed-gpu.md#tensorflow)
2929

30-
For machine learning models that don't require distributed training, see [train models with Azure Machine Learning](concept-train-machine-learning-model.md#python-sdk) for the different ways to train models using the Python SDK.
30+
For machine learning models that don't require distributed training, see [Train models with Azure Machine Learning](concept-train-machine-learning-model.md#python-sdk) for different ways to train models using the Python SDK.
3131

3232
## Data parallelism
3333

3434
Data parallelism is the easiest to implement of the two distributed training approaches, and is sufficient for most use cases.
3535

36-
In this approach, the data is divided into partitions, where the number of partitions is equal to the total number of available nodes, in the compute cluster or [serverless compute](./how-to-use-serverless-compute.md). The model is copied in each of these worker nodes, and each node operates on its own subset of the data. Keep in mind that each node must have the capacity to support the model that's being trained, that is, the model has to entirely fit on each node.
36+
In this approach, the data is divided into partitions, where the number of partitions is equal to the total number of available nodes, in the compute cluster or [serverless compute](./how-to-use-serverless-compute.md). The model is copied in each of these worker nodes, and each node operates on its own subset of the data. Keep in mind that each node must have the capacity to support the model that's being trained, that is, the entire model has to fit on each node.
3737

38-
The following diagram provides a visual demonstration of this approach.
38+
The following diagram shows this approach.
3939

4040
:::image type="content" source="media/concept-distributed-training/distributed-training.svg" alt-text="Diagram of data parrallelism showing the model copied into worker nodes.":::
4141

0 commit comments

Comments
 (0)