You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/synapse-analytics/machine-learning/tutorial-horovod-pytorch.md
+6-6Lines changed: 6 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -21,7 +21,7 @@ Within Azure Synapse Analytics, users can quickly get started with Horovod using
21
21
- Create a GPU-enabled Apache Spark pool in your Azure Synapse Analytics workspace. For details, see [Create a GPU-enabled Apache Spark pool in Azure Synapse](../spark/apache-spark-gpu-concept.md). For this tutorial, we suggest using the GPU-Large cluster size with 3 nodes.
22
22
23
23
> [!NOTE]
24
-
> The Preview for Azure Synapse GPU-enabled pools has now been deprecated.
24
+
> The Preview for Azure Synapse GPU-enabled pools is deprecated.
25
25
26
26
## Configure the Apache Spark session
27
27
@@ -46,7 +46,7 @@ In the example, you can see how the Spark configurations can be passed with the
46
46
}
47
47
```
48
48
49
-
For this tutorial, we will use the following configurations:
49
+
For this tutorial, we'll use the following configurations:
50
50
51
51
```python
52
52
@@ -97,7 +97,7 @@ from azure.synapse.ml.horovodutils import AdlsStore
97
97
98
98
## Connect to alternative storage account
99
99
100
-
We need the Azure Data Lake Storage (ADLS) account for storing intermediate and model data. If you are using an alternative storage account, be sure to set up the [linked service](../../data-factory/concepts-linked-services.md) to automatically authenticate and read from the account. In addition, you need to modify the following properties: ```remote_url```, ```account_name```, and ```linked_service_name```.
100
+
We need the Azure Data Lake Storage (ADLS) account for storing intermediate and model data. If you're using an alternative storage account, be sure to set up the [linked service](../../data-factory/concepts-linked-services.md) to automatically authenticate and read from the account. In addition, you need to modify the following properties: ```remote_url```, ```account_name```, and ```linked_service_name```.
101
101
102
102
```python
103
103
num_proc =3# equal to numExecutors
@@ -127,7 +127,7 @@ print(adls_store_path)
127
127
128
128
## Prepare dataset
129
129
130
-
Next, we will prepare the dataset for training. In this tutorial, we will use the MNIST dataset from [Azure Open Datasets](/azure/open-datasets/dataset-mnist?tabs=azureml-opendatasets).
130
+
Next, we'll prepare the dataset for training. In this tutorial, we'll use the MNIST dataset from [Azure Open Datasets](/azure/open-datasets/dataset-mnist?tabs=azureml-opendatasets).
131
131
132
132
```python
133
133
# Initialize SparkSession
@@ -148,7 +148,7 @@ mnist_df.head()
148
148
149
149
## Process data with Apache Spark
150
150
151
-
Now, we will create an Apache Spark dataframe. This dataframe will be used with the ```HorovodEstimator``` for training.
151
+
Now, we'll create an Apache Spark dataframe. This dataframe will be used with the ```HorovodEstimator``` for training.
152
152
153
153
```python
154
154
# Create Spark DataFrame for training
@@ -167,7 +167,7 @@ train_df.count()
167
167
168
168
## Define DNN model
169
169
170
-
Once we are finished processing our dataset, we can now define our PyTorch model. The same code could also be used to train a single-node PyTorch model.
170
+
Once we're finished processing our dataset, we can now define our PyTorch model. The same code could also be used to train a single-node PyTorch model.
171
171
172
172
```python
173
173
# Define the PyTorch model without any Horovod-specific parameters
0 commit comments