Skip to content

Commit 5a7781e

Browse files
Merge pull request #247074 from xiaoharper/patch-6
fix github issues
2 parents 08b4e51 + 33897f9 commit 5a7781e

File tree

1 file changed

+6
-6
lines changed

1 file changed

+6
-6
lines changed

articles/machine-learning/v1/how-to-move-data-in-out-of-pipelines.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -76,23 +76,23 @@ For more options on creating datasets with different options and from different
7676

7777
To pass the dataset's path to your script, use the `Dataset` object's `as_named_input()` method. You can either pass the resulting `DatasetConsumptionConfig` object to your script as an argument or, by using the `inputs` argument to your pipeline script, you can retrieve the dataset using `Run.get_context().input_datasets[]`.
7878

79-
Once you've created a named input, you can choose its access mode: `as_mount()` or `as_download()`. If your script processes all the files in your dataset and the disk on your compute resource is large enough for the dataset, the download access mode is the better choice. The download access mode avoids the overhead of streaming the data at runtime. If your script accesses a subset of the dataset or it's too large for your compute, use the mount access mode. For more information, read [Mount vs. Download](how-to-train-with-datasets.md#mount-vs-download)
79+
Once you've created a named input, you can choose its access mode(for FileDataset only): `as_mount()` or `as_download()`. If your script processes all the files in your dataset and the disk on your compute resource is large enough for the dataset, the download access mode is the better choice. The download access mode avoids the overhead of streaming the data at runtime. If your script accesses a subset of the dataset or it's too large for your compute, use the mount access mode. For more information, read [Mount vs. Download](how-to-train-with-datasets.md#mount-vs-download)
8080

8181
To pass a dataset to your pipeline step:
8282

8383
1. Use `TabularDataset.as_named_input()` or `FileDataset.as_named_input()` (no 's' at end) to create a `DatasetConsumptionConfig` object
84-
1. Use `as_mount()` or `as_download()` to set the access mode
84+
1. **For `FileDataset` only:**. Use `as_mount()` or `as_download()` to set the access mode. TabularDataset does not suppport set access mode.
8585
1. Pass the datasets to your pipeline steps using either the `arguments` or the `inputs` argument
8686

87-
The following snippet shows the common pattern of combining these steps within the `PythonScriptStep` constructor:
87+
The following snippet shows the common pattern of combining these steps within the `PythonScriptStep` constructor, using iris_dataset (TabularDataset):
8888

8989
```python
9090

9191
train_step = PythonScriptStep(
9292
name="train_data",
9393
script_name="train.py",
9494
compute_target=cluster,
95-
inputs=[iris_dataset.as_named_input('iris').as_mount()]
95+
inputs=[iris_dataset.as_named_input('iris')]
9696
)
9797
```
9898

@@ -111,7 +111,7 @@ train_step = PythonScriptStep(
111111
name="train_data",
112112
script_name="train.py",
113113
compute_target=cluster,
114-
inputs=[train.as_named_input('train').as_download(), test.as_named_input('test').as_download()]
114+
inputs=[train.as_named_input('train'), test.as_named_input('test')]
115115
)
116116
```
117117

@@ -266,4 +266,4 @@ For more information, see [Plan and manage costs for Azure Machine Learning](../
266266
## Next steps
267267

268268
* [Create an Azure Machine Learning dataset](how-to-create-register-datasets.md)
269-
* [Create and run machine learning pipelines with Azure Machine Learning SDK](how-to-create-machine-learning-pipelines.md)
269+
* [Create and run machine learning pipelines with Azure Machine Learning SDK](how-to-create-machine-learning-pipelines.md)

0 commit comments

Comments
 (0)