Skip to content

Commit c387339

Browse files
authored
Merge pull request #106301 from likebupt/update-0303
Update 0303
2 parents d349287 + cf2bd76 commit c387339

File tree

7 files changed

+61
-21
lines changed

7 files changed

+61
-21
lines changed

articles/machine-learning/algorithm-module-reference/create-python-model.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ Use of this module requires intermediate or expert knowledge of Python. The modu
3030

3131
This article will show how to use the **Create Python Model** with a simple pipeline. Below is the graph of the pipeline.
3232

33-
![create-python-model](./media/module/aml-create-python-model.png)
33+
![create-python-model](./media/module/create-python-model.png)
3434

3535
1. Click **Create Python Model**, edit the script to implement your modeling or data management process. You can base the model on any learner that is included in a Python package in the Azure Machine Learning environment.
3636

articles/machine-learning/algorithm-module-reference/enter-data-manually.md

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -84,11 +84,9 @@ This module can be helpful in scenarios such as these:
8484
|0.00016|0.004|0.999961|0.00784|1|
8585
|0|0.004|0.999955|0.008615|1|
8686
87-
4. Press ENTER after each row, to start a new line.
88-
89-
**Be sure to press ENTER after the final row.**
87+
4. Press ENTER after each row, to start a new line.
9088
91-
If you press ENTER multiple times to add multiple empty trailing rows, the final empty row is removed trimmed, but other empty rows are treated as missing values.
89+
If you press ENTER multiple times to add multiple empty trailing rows, the empty rows will be removed trimmed.
9290
9391
If you create rows with missing values, you can always filter them out later.
9492
66.5 KB
Loading
41 KB
Loading

articles/machine-learning/algorithm-module-reference/partition-and-sample.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -149,9 +149,9 @@ This option is used when you have divided a dataset into multiple partitions and
149149

150150
5. If you are working with multiple partitions, you must add additional instances of the **Partition and Sample** module to handle each partition.
151151

152-
For example, let's say previously partitioned patients into four folds using age. To work with each individual fold, you need four copies of the **Partition and Sample** module, and in each, you select a different fold, as shown below. It's not correct to use the **Assign to Folds** output directly.
152+
For example, the **Partition and Sample** module in the second row is set to **Assign to Folds**, and the modules in the third row is set to **Pick Fold**.
153153

154-
[![Partition and sample](./media/partition-and-sample/partition-and-sample.png)](./media/partition-and-sample/partition-and-sample-lg.png#lightbox)
154+
![Partition and sample](./media/module/partition-and-sample.png)
155155

156156
5. Run the pipeline.
157157

articles/machine-learning/algorithm-module-reference/split-data.md

Lines changed: 52 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -79,35 +79,75 @@ This module is particularly useful when you need to separate data into training
7979

8080
Based on the regular expression you provide, the dataset is divided into two sets of rows: rows with values that match the expression and all remaining rows.
8181

82+
The following examples demonstrate how to divide a dataset using the **Regular Expression** option.
83+
84+
### Single whole word
85+
86+
This example puts into the first dataset all rows that contain the text `Gryphon` in the column `Text`, and puts other rows into the second output of **Split Data**:
87+
88+
```text
89+
\"Text" Gryphon
90+
```
91+
92+
### Substring
93+
94+
This example looks for the specified string in any position within the second column of the dataset, denoted here by the index value of 1. The match is case-sensitive.
95+
96+
```text
97+
(\1) ^[a-f]
98+
```
99+
100+
The first result dataset contains all rows where the index column begins with one of these characters: `a`, `b`, `c`, `d`, `e`, `f`. All other rows are directed to the second output.
101+
82102
## Relative expression split.
83103

84104
1. Add the [Split Data](./split-data.md) module to your pipeline, and connect it as input to the dataset you want to split.
85105

86106
2. For **Splitting mode**, select **relative expression split**.
87107

88-
3. In the **Relational expression** text box, type an expression that performs a comparison operation, on a single column:
108+
3. In the **Relational expression** text box, type an expression that performs a comparison operation on a single column:
109+
110+
For the **Numeric column**:
111+
- The column contains numbers of any numeric data type, including date and time data types.
112+
- The expression can reference a maximum of one column name.
113+
- Use the ampersand character `&` for the AND operation. Use the pipe character `|` for the OR operation.
114+
- The following operators are supported: `<`, `>`, `<=`, `>=`, `==`, `!=`.
115+
- You cannot group operations by using `(` and `)`.
116+
117+
For the **String column**:
118+
- The following operators are supported: `==`, `!=`.
89119

120+
4. Run the pipeline.
90121

91-
- Numeric column:
92-
- The column contains numbers of any numeric data type, including date/time data types.
122+
The expression divides the dataset into two sets of rows: rows with values that meet the condition, and all remaining rows.
93123

94-
- The expression can reference a maximum of one column name.
124+
The following examples demonstrate how to divide a dataset using the **Relative Expression** option in the **Split Data** module:
95125

96-
- Use the ampersand character (&) for the AND operation and use the pipe character (|) for the OR operation.
126+
### Using calendar year
97127

98-
- The following operators are supported: `<`, `>`, `<=`, `>=`, `==`, `!=`
128+
A common scenario is to divide a dataset by years. The following expression selects all rows where the values in the column `Year` are greater than `2010`.
99129

100-
- You cannot group operations by using `(` and `)`.
130+
```text
131+
\"Year" > 2010
132+
```
101133

102-
- String column:
103-
- The following operators are supported: `==`, `!=`
134+
The date expression must account for all date parts that are included in the data column, and the format of dates in the data column must be consistent.
104135

136+
For example, in a date column using the format `mmddyyyy`, the expression should be something like this:
105137

138+
```text
139+
\"Date" > 1/1/2010
140+
```
106141

107-
4. Run the pipeline.
142+
### Using column indices
143+
144+
The following expression demonstrates how you can use the column index to select all rows in the first column of the dataset that contain values less than or equal to 30, but not equal to 20.
145+
146+
```text
147+
(\0)<=30 & !=20
148+
```
108149

109-
The expression divides the dataset into two sets of rows: rows with values that meet the condition, and all remaining rows.
110150

111151
## Next steps
112152

113-
See the [set of modules available](module-reference.md) to Azure Machine Learning.
153+
See the [set of modules available](module-reference.md) to Azure Machine Learning.

articles/machine-learning/how-to-retrain-designer.md

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -114,15 +114,17 @@ Use the following steps to submit a pipeline endpoint run from the designer:
114114

115115
1. Select the pipeline you want to run.
116116

117-
1. Select **Run**.
117+
1. Select **Submit**.
118118

119119
1. In the setup dialog, you can specify a new input data path value, which points to your new dataset.
120120

121121
![Screenshot showing how to set up a parameterized pipeline run in the designer](./media/how-to-retrain-designer/published-pipeline-run.png)
122122

123123
### Submit runs with code
124124

125-
There are multiple ways to access your REST endpoint programatically depending on your development environment. You can find code samples that show you how to submit pipeline runs with parameters in the **Consume** tab of your pipeline.
125+
You can find the REST endpoint of a published pipeline in the overview panel. By calling the endpoint, you can retrain the published pipeline.
126+
127+
To make a REST call, you will need an OAuth 2.0 bearer-type authentication header. See the following [tutorial section](tutorial-pipeline-batch-scoring-classification.md#publish-and-run-from-a-rest-endpoint) for more detail on setting up authentication to your workspace and making a parameterized REST call.
126128

127129
## Next steps
128130

0 commit comments

Comments
 (0)