Skip to content

Commit f57b640

Browse files
committed
team feedback
1 parent cfb52da commit f57b640

File tree

1 file changed

+7
-6
lines changed

1 file changed

+7
-6
lines changed

articles/machine-learning/concept-distributed-training.md

Lines changed: 7 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
---
22
title: What is distributed training?
33
titleSuffix: Azure Machine Learning
4-
description: Distributed training refers to the ability to accelerate model training by sharing and parallelizing data loads and training tasks across multiple GPUs.
4+
description: Learn about distributed training and how Azure Machine Learning supports it.
55
services: machine-learning
66
ms.service: machine-learning
77
author: nibaccam
@@ -13,18 +13,19 @@ ms.date: 03/27/2020
1313

1414
# Distributed training with Azure Machine Learning
1515

16-
In distributed training the work load to train a model is split up and shared among multiple mini processors, called worker nodes. These worker nodes work in parallel to speed up model training. Distrbuted training is well suited for compute and time intensive tasks, like [deep learning](concept-deep-learning-vs-machine-learning.md) for training deep neural networks.
16+
In this article, you learn about distributed training and how Azure Machine Learning supports it.
1717

18-
## Distributed training in Azure Machine Learning
18+
In distributed training the workload to train a model is split up and shared among multiple mini processors, called worker nodes. These worker nodes work in parallel to speed up model training. Distributed training is well suited for compute and time intensive tasks, like [deep learning](concept-deep-learning-vs-machine-learning.md) for training deep neural networks.
1919

20-
There are two main types of distributed training: [data parallelism](#data-parallelism) and [model parallelism](#model-parallelism).
21-
The [Azure Machine Learning SDK in Python](https://docs.microsoft.com/python/api/overview/azure/ml/intro?view=azure-ml-py) supports integrations with popular deep learning frameworks, PyTorch and TensorFlow. Both frameworks employ data parallelism for distributed training, and leverage [horovod](https://horovod.readthedocs.io/en/latest/summary_include.html) for optimizing compute speeds.
20+
## Deep learning and distributed training
21+
22+
There are two main types of distributed training: [data parallelism](#data-parallelism) and [model parallelism](#model-parallelism). For distributed training, the [Azure Machine Learning SDK in Python](https://docs.microsoft.com/python/api/overview/azure/ml/intro?view=azure-ml-py) supports integrations with popular deep learning frameworks, PyTorch and TensorFlow. Both frameworks employ data parallelism for distributed training, and leverage [horovod](https://horovod.readthedocs.io/en/latest/summary_include.html) for optimizing compute speeds.
2223

2324
* [Distributed training with PyTorch](how-to-train-pytorch.md#distributed-training)
2425

2526
* [Distributed training with TensorFlow](how-to-train-tensorflow.md#distributed-training)
2627

27-
For training traditional ML models, see [Azure Machine Learning SDK for Python](concept-train-machine-learning-model.md#python-sdk) for the different ways to train models using the Python SDK.
28+
For training traditional ML models, see [train models with Azure Machine Learning](concept-train-machine-learning-model.md#python-sdk) for the different ways to train models using the Python SDK.
2829

2930
## Data parallelism
3031

0 commit comments

Comments
 (0)