You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The [datastore](how-to-access-data.md) is a place where data can be stored and accessed by mounting or copying the data to the compute target. Each workspace provides a default datastore. Upload the data and training scripts to the datastore so that they can be easily accessed during training.
80
+
### Create a file dataset
90
81
91
-
1. Download the MNIST dataset locally.
82
+
A `FileDataset` object references one or multiple files in your workspace datastore or public urls. The files can be of any format, and the class provides you with the ability to download or mount the files to your compute. By creating a `FileDataset`, you create a reference to the data source location. If you applied any transformations to the data set, they will be stored in the data set as well. The data remains in its existing location, so no extra storage cost is incurred. See the [how-to](https://docs.microsoft.com/en-us/azure/machine-learning/service/how-to-create-register-datasets) guide on the `Dataset` package for more information.
1. Upload the Keras training script, `keras_mnist.py`, andthe helper file, `utils.py`.
96
+
Use the `register()` method to register the data set to your workspace so they can be shared with others, reused across various experiments, and referred to by name in your training script.
110
97
111
-
```Python
112
-
shutil.copy('./keras_mnist.py', script_folder)
113
-
shutil.copy('./utils.py', script_folder)
114
-
```
98
+
```python
99
+
dataset = dataset.register(workspace=ws,
100
+
name='mnist dataset',
101
+
description='training and test dataset',
102
+
create_new_version=True)
103
+
```
115
104
116
105
## Create a compute target
117
106
@@ -139,11 +128,22 @@ For more information on compute targets, see the [what is a compute target](conc
139
128
140
129
The [TensorFlow estimator](https://docs.microsoft.com/python/api/azureml-train-core/azureml.train.dnn.tensorflow?view=azure-ml-py) provides a simple way of launching TensorFlow training jobs on compute target. Since Keras runs on top of TensorFlow, you can use the TensorFlow estimator and import the Keras library using the `pip_packages` argument.
141
130
131
+
First get the data from the workspace datastore using the `Dataset` class.
The TensorFlow estimator is implemented through the generic [`estimator`](https://docs.microsoft.com//python/api/azureml-train-core/azureml.train.estimator.estimator?view=azure-ml-py) class, which can be used to support any framework. Additionally, create a dictionary `script_params` that contains the DNN hyperparameter settings. For more information about training models using the generic estimator, see [train models with Azure Machine Learning using estimator](how-to-train-ml-models.md)
0 commit comments