Skip to content

Commit dd19916

Browse files
authored
add deepspeed example link
add example link to how-to-train-distributed-gpu.md
1 parent 4bfdba8 commit dd19916

File tree

1 file changed

+4
-0
lines changed

1 file changed

+4
-0
lines changed

articles/machine-learning/how-to-train-distributed-gpu.md

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -130,6 +130,10 @@ Azure ML will set the `MASTER_ADDR`, `MASTER_PORT`, `WORLD_SIZE`, and `NODE_RANK
130130

131131
You can use a [curated environment](resource-curated-environments.md#azure-container-for-pytorch-acpt-preview) for an out of the box environment with the latest state of art technologies including `DeepSpeed`, `ORT`, `MSSCCL`, and `Pytorch` for your DeepSpeed training jobs.
132132

133+
### DeepSpeed example
134+
135+
- For DeepSpeed training and autotuning examples, see [here](https://github.com/Azure/azureml-examples/tree/main/cli/jobs/deepspeed).
136+
133137
## TensorFlow
134138

135139
If you're using [native distributed TensorFlow](https://www.tensorflow.org/guide/distributed_training) in your training code, such as TensorFlow 2.x's `tf.distribute.Strategy` API, you can launch the distributed job via Azure ML using `distribution` parameters or the `TensorFlowDistribution` object.

0 commit comments

Comments
 (0)