Skip to content

Commit 6e149b1

Browse files
authored
Merge pull request #105253 from likebupt/update-module-articles
update import/export data module
2 parents c7c101a + adacfb1 commit 6e149b1

File tree

2 files changed

+16
-7
lines changed

2 files changed

+16
-7
lines changed

articles/machine-learning/algorithm-module-reference/export-data.md

Lines changed: 10 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -9,13 +9,13 @@ ms.topic: reference
99

1010
author: likebupt
1111
ms.author: keli19
12-
ms.date: 10/22/2019
12+
ms.date: 02/22/2020
1313
---
1414
# Export Data module
1515

1616
This article describes a module in Azure Machine Learning designer (preview).
1717

18-
Use this module to save results, intermediate data, and working data from your pipelines into cloud storage destinations outside Azure Machine Learning.
18+
Use this module to save results, intermediate data, and working data from your pipelines into cloud storage destinations.
1919

2020
This module supports exporting your data to the following cloud data services:
2121

@@ -24,7 +24,7 @@ This module supports exporting your data to the following cloud data services:
2424
- Azure Data Lake
2525
- Azure Data Lake Gen2
2626

27-
Before exporting your data, you need to first register a datastore in your Azure Machine Learning workspace first. For more information, see [Access data in Azure storage services](../how-to-access-data.md).
27+
Before exporting your data, you need to first register a datastore in your Azure Machine Learning workspace. For more information, see [Access data in Azure storage services](../how-to-access-data.md).
2828

2929
## How to configure Export Data
3030

@@ -36,7 +36,13 @@ Before exporting your data, you need to first register a datastore in your Azure
3636

3737
1. For **Datastore**, select an existing datastore from the dropdown list. You can also create a new datastore. Check how by visiting [Access data in Azure storage services](../how-to-access-data.md).
3838

39-
1. Define the path in the datastore to write the data to.
39+
1. The checkbox, **Regenerate output**, decides whether to execute the module to regenerate output at running time.
40+
41+
It's by default unselected, which means if the module has been executed with the same parameters previously, the system will reuse the output from last run to reduce run time.
42+
43+
If it is selected, the system will execute the module again to regenerate output.
44+
45+
1. Define the path in the datastore where the data is. The path is a relative path. The empty paths or a URL paths are not allowed.
4046

4147

4248
1. For **File format**, select the format in which data should be stored.

articles/machine-learning/algorithm-module-reference/import-data.md

Lines changed: 6 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -43,9 +43,7 @@ If your source data changes, you can refresh the dataset and add new data by rer
4343

4444
1. Add the **Import Data** module to your pipeline. You can find this module in the **Data Input and Output** category in the designer.
4545

46-
1. Click **Launch Data Import Wizard** to configure the data source using a wizard.
47-
48-
The wizard gets the account name and credentials, and help you configure other options. If you are editing an existing configuration, it loads the current values first.
46+
1. Select the module to open the right pane.
4947

5048
1. Select **Data source**, and choose the data source type. It could be HTTP or datastore.
5149

@@ -56,6 +54,11 @@ If your source data changes, you can refresh the dataset and add new data by rer
5654

5755
![import-data-preview](media/module/import-data.png)
5856

57+
1. The checkbox, **Regenerate output**, decides whether to execute the module to regenerate output at running time.
58+
59+
It's by default unselected, which means if the module has been executed with the same parameters previously, the system will reuse the output from last run to reduce run time.
60+
61+
If it is selected, the system will execute the module again to regenerate output. So select this option when underlying data in storage is updated, it can help to get the latest data.
5962

6063

6164
1. Run the pipeline.

0 commit comments

Comments
 (0)