You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
1. Click **Create Python Model**, edit the script to implement your modeling or data management process. You can base the model on any learner that is included in a Python package in the Azure Machine Learning environment.
Copy file name to clipboardExpand all lines: articles/machine-learning/algorithm-module-reference/enter-data-manually.md
+2-4Lines changed: 2 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -84,11 +84,9 @@ This module can be helpful in scenarios such as these:
84
84
|0.00016|0.004|0.999961|0.00784|1|
85
85
|0|0.004|0.999955|0.008615|1|
86
86
87
-
4. Press ENTER after each row, to start a new line.
88
-
89
-
**Be sure to press ENTER after the final row.**
87
+
4. Press ENTER after each row, to start a new line.
90
88
91
-
If you press ENTER multiple times to add multiple empty trailing rows, the final empty row is removed trimmed, but other empty rows are treated as missing values.
89
+
If you press ENTER multiple times to add multiple empty trailing rows, the empty rows will be removed trimmed.
92
90
93
91
If you create rows with missing values, you can always filter them out later.
Copy file name to clipboardExpand all lines: articles/machine-learning/algorithm-module-reference/partition-and-sample.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -149,9 +149,9 @@ This option is used when you have divided a dataset into multiple partitions and
149
149
150
150
5. If you are working with multiple partitions, you must add additional instances of the **Partition and Sample** module to handle each partition.
151
151
152
-
For example, let's say previously partitioned patients into four folds using age. To work with each individual fold, you need four copies of the **Partition and Sample** module, and in each, you select a different fold, as shown below. It's not correct to use the **Assign to Folds** output directly.
152
+
For example, the **Partition and Sample** modulein the second row is set to **Assign to Folds**, and the modules in the third row is set to **Pick Fold**.
153
153
154
-
[](./media/partition-and-sample/partition-and-sample-lg.png#lightbox)
154
+

Copy file name to clipboardExpand all lines: articles/machine-learning/algorithm-module-reference/split-data.md
+52-12Lines changed: 52 additions & 12 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -79,35 +79,75 @@ This module is particularly useful when you need to separate data into training
79
79
80
80
Based on the regular expression you provide, the dataset is divided into two sets of rows: rows with values that match the expression and all remaining rows.
81
81
82
+
The following examples demonstrate how to divide a dataset using the **Regular Expression** option.
83
+
84
+
### Single whole word
85
+
86
+
This example puts into the first dataset all rows that contain the text `Gryphon` in the column `Text`, and puts other rows into the second output of **Split Data**:
87
+
88
+
```text
89
+
\"Text" Gryphon
90
+
```
91
+
92
+
### Substring
93
+
94
+
This example looks for the specified string in any position within the second column of the dataset, denoted here by the index value of 1. The match is case-sensitive.
95
+
96
+
```text
97
+
(\1) ^[a-f]
98
+
```
99
+
100
+
The first result dataset contains all rows where the index column begins with one of these characters: `a`, `b`, `c`, `d`, `e`, `f`. All other rows are directed to the second output.
101
+
82
102
## Relative expression split.
83
103
84
104
1. Add the [Split Data](./split-data.md) module to your pipeline, and connect it as input to the dataset you want to split.
85
105
86
106
2. For **Splitting mode**, select **relative expression split**.
87
107
88
-
3. In the **Relational expression** text box, type an expression that performs a comparison operation, on a single column:
108
+
3. In the **Relational expression** text box, type an expression that performs a comparison operation on a single column:
109
+
110
+
For the **Numeric column**:
111
+
- The column contains numbers of any numeric data type, including date and time data types.
112
+
- The expression can reference a maximum of one column name.
113
+
- Use the ampersand character `&` for the AND operation. Use the pipe character `|` for the OR operation.
114
+
- The following operators are supported: `<`, `>`, `<=`, `>=`, `==`, `!=`.
115
+
- You cannot group operations by using `(` and `)`.
116
+
117
+
For the **String column**:
118
+
- The following operators are supported: `==`, `!=`.
89
119
120
+
4. Run the pipeline.
90
121
91
-
- Numeric column:
92
-
- The column contains numbers of any numeric data type, including date/time data types.
122
+
The expression divides the dataset into two sets of rows: rows with values that meet the condition, and all remaining rows.
93
123
94
-
- The expression can reference a maximum of one column name.
124
+
The following examples demonstrate how to divide a dataset using the **Relative Expression** option in the **Split Data** module:
95
125
96
-
- Use the ampersand character (&) for the AND operation and use the pipe character (|) for the OR operation.
126
+
### Using calendar year
97
127
98
-
-The following operators are supported: `<`, `>`, `<=`, `>=`, `==`, `!=`
128
+
A common scenario is to divide a dataset by years. The following expression selects all rows where the values in the column `Year` are greater than `2010`.
99
129
100
-
- You cannot group operations by using `(` and `)`.
130
+
```text
131
+
\"Year" > 2010
132
+
```
101
133
102
-
- String column:
103
-
- The following operators are supported: `==`, `!=`
134
+
The date expression must account for all date parts that are included in the data column, and the format of dates in the data column must be consistent.
104
135
136
+
For example, in a date column using the format `mmddyyyy`, the expression should be something like this:
105
137
138
+
```text
139
+
\"Date" > 1/1/2010
140
+
```
106
141
107
-
4. Run the pipeline.
142
+
### Using column indices
143
+
144
+
The following expression demonstrates how you can use the column index to select all rows in the first column of the dataset that contain values less than or equal to 30, but not equal to 20.
145
+
146
+
```text
147
+
(\0)<=30 & !=20
148
+
```
108
149
109
-
The expression divides the dataset into two sets of rows: rows with values that meet the condition, and all remaining rows.
110
150
111
151
## Next steps
112
152
113
-
See the [set of modules available](module-reference.md) to Azure Machine Learning.
153
+
See the [set of modules available](module-reference.md) to Azure Machine Learning.
Copy file name to clipboardExpand all lines: articles/machine-learning/how-to-retrain-designer.md
+4-2Lines changed: 4 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -114,15 +114,17 @@ Use the following steps to submit a pipeline endpoint run from the designer:
114
114
115
115
1. Select the pipeline you want to run.
116
116
117
-
1. Select **Run**.
117
+
1. Select **Submit**.
118
118
119
119
1. In the setup dialog, you can specify a new input data path value, which points to your new dataset.
120
120
121
121

122
122
123
123
### Submit runs with code
124
124
125
-
There are multiple ways to access your REST endpoint programatically depending on your development environment. You can find code samples that show you how to submit pipeline runs with parameters in the **Consume** tab of your pipeline.
125
+
You can find the REST endpoint of a published pipeline in the overview panel. By calling the endpoint, you can retrain the published pipeline.
126
+
127
+
To make a REST call, you will need an OAuth 2.0 bearer-type authentication header. See the following [tutorial section](tutorial-pipeline-batch-scoring-classification.md#publish-and-run-from-a-rest-endpoint) for more detail on setting up authentication to your workspace and making a parameterized REST call.
0 commit comments