You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/machine-learning/service/how-to-ui-sample-classification-predict-churn.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -34,7 +34,7 @@ Here's the completed graph for this experiment:
34
34
35
35
## Data
36
36
37
-
The data for this pipeline is from KDD Cup 2009. It has 50,000 rows and 230 feature columns. The task is to predict churn, appetency, and up-selling for customers who use these features. For more information about the data and the task, see the [KDD website](https://www.kdd.org/kdd-cup/view/kdd-cup-2009).
37
+
The data for this experiment is from KDD Cup 2009. It has 50,000 rows and 230 feature columns. The task is to predict churn, appetency, and up-selling for customers who use these features. For more information about the data and the task, see the [KDD website](https://www.kdd.org/kdd-cup/view/kdd-cup-2009).
Copy file name to clipboardExpand all lines: articles/machine-learning/service/how-to-ui-sample-classification-predict-credit-risk-basic.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -18,7 +18,7 @@ Learn how to build a machine learning classifier without writing a single line o
18
18
19
19
Because the question is answering "Which one?" this is called a classification problem. However, you can apply the same fundamental process to tackle any type of machine learning problem whether it be regression, classification, clustering, and so on.
20
20
21
-
Here's the final pipeline graph for this sample:
21
+
Here's the final experiment graph for this sample:
22
22
23
23

24
24
@@ -33,7 +33,7 @@ Here's the final pipeline graph for this sample:
provides an advanced pipeline that solves the same problem as this pipeline. It shows how to perform *cost sensitive* classification by using an **Execute Python Script** module and compare the performance of two binary classification algorithms. Refer to it if you want to learn more about how to build classification pipelines.
36
+
provides an advanced experiment that solves the same problem as this experiment. It shows how to perform *cost sensitive* classification by using an **Execute Python Script** module and compare the performance of two binary classification algorithms. Refer to it if you want to learn more about how to build classification pipelines.
37
37
38
38
## Data
39
39
@@ -42,7 +42,7 @@ The dataset contains 1,000 samples with 20 features and 1 label. Each sample rep
42
42
43
43
## Experiment summary
44
44
45
-
Follow these steps to create the pipeline:
45
+
Follow these steps to create the experiment:
46
46
47
47
1. Drag the German Credit Card UCI Data dataset module into the experiment's canvas.
48
48
1. Add an **Edit Metadata** module so we can add meaningful names for each column.
In the evaluation stage of the pipeline, you compute the accuracy of each of the four models. For this pipeline, use **Evaluate Model** to compare examples that have the same misclassification cost.
110
+
In the evaluation stage of the experiment, you compute the accuracy of each of the four models. For this experiment, use **Evaluate Model** to compare examples that have the same misclassification cost.
111
111
112
112
The **Evaluate Model** module can compute the performance metrics for as many as two scored models. So you can use one instance of **Evaluate Model** to evaluate the two SVM models and another instance of **Evaluate Model** to evaluate the two Boosted Decision Tree models.
Copy file name to clipboardExpand all lines: articles/machine-learning/service/how-to-ui-sample-classification-predict-flight-delay.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -18,9 +18,9 @@ This experiment uses historical flight and weather data to predict if a schedule
18
18
19
19
This problem can be approached as a classification problem, predicting two classes -- delayed, or on time. To build a classifier, this model using a large number of examples from historic flight data.
20
20
21
-
Here's the final pipeline graph for this sample:
21
+
Here's the final experiment graph for this sample:
22
22
23
-
[](media/ui-sample-classification-predict-credit-risk-cost-sensitive/graph.png#lightbox)
23
+
[](media/ui-sample-classification-predict-credit-risk-cost-sensitive/graph.png#lightbox)
Copy file name to clipboardExpand all lines: articles/machine-learning/service/how-to-ui-sample-regression-predict-automobile-price-basic.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -16,7 +16,7 @@ ms.date: 05/10/2019
16
16
17
17
Learn how to build a machine learning regression model without writing a single line of code using the visual interface.
18
18
19
-
This pipeline trains a **decision forest regressor** to predict a car's price based on technical features such as make, model, horsepower, and size. Because you're trying to answer the question "How much?" this is called a regression problem. However, you can apply the same fundamental steps in this example to tackle any type of machine learning problem whether it be regression, classification, clustering, and so on.
19
+
This experiment trains a **decision forest regressor** to predict a car's price based on technical features such as make, model, horsepower, and size. Because you're trying to answer the question "How much?" this is called a regression problem. However, you can apply the same fundamental steps in this example to tackle any type of machine learning problem whether it be regression, classification, clustering, and so on.
20
20
21
21
The fundamental steps of a training machine learning model are:
22
22
@@ -39,7 +39,7 @@ Here's the final, completed graph of the experiment we'll be working on. We'll p
39
39
40
40
## Get the data
41
41
42
-
This sample uses the **Automobile price data (Raw)** dataset, which is from the UCI Machine Learning Repository. The dataset contains 26 columns that contain information about automobiles, including make, model, price, vehicle features (like the number of cylinders), MPG, and an insurance risk score. The goal of this pipeline is to predict the price of the car.
42
+
This sample uses the **Automobile price data (Raw)** dataset, which is from the UCI Machine Learning Repository. The dataset contains 26 columns that contain information about automobiles, including make, model, price, vehicle features (like the number of cylinders), MPG, and an insurance risk score. The goal of this experiment is to predict the price of the car.
Copy file name to clipboardExpand all lines: articles/machine-learning/service/how-to-ui-sample-regression-predict-automobile-price-compare-algorithms.md
+5-5Lines changed: 5 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -16,7 +16,7 @@ ms.date: 05/10/2019
16
16
17
17
Learn how to build a complex machine learning experiment without writing a single line of code using the visual interface. This sample trains and compares multiple regression models to predict a car's price based on its technical features. We'll provide the rationale for the choices made in this experiment so you can tackle your own machine learning problems.
18
18
19
-
If you're just getting started with machine learning, take a look at the [basic version](how-to-ui-sample-regression-predict-automobile-price-basic.md) of this pipeline.
19
+
If you're just getting started with machine learning, take a look at the [basic version](how-to-ui-sample-regression-predict-automobile-price-basic.md) of this experiment.
20
20
21
21
Here's the completed graph for this experiment:
22
22
@@ -32,7 +32,7 @@ Here's the completed graph for this experiment:
32
32
33
33
## Experiment summary
34
34
35
-
Use following steps to build the machine learning pipeline:
35
+
Use following steps to build the machine learning experiment:
36
36
37
37
1. Get the data.
38
38
1. Pre-process the data.
@@ -57,9 +57,9 @@ Machine learning problems vary. Common machine learning tasks include classifica
57
57
58
58
Because the goal of this experiment is to predict automobile prices, and because the label column (price) contains real numbers, a regression model is a good choice. Considering that the number of features is relatively small (less than 100) and these features aren't sparse, the decision boundary is likely to be nonlinear.
59
59
60
-
Because the goal of this pipeline is to predict automobile prices, and because the label column (price) contains real numbers, a regression model is a good choice. Considering that the number of features is relatively small (less than 100) and these features aren't sparse, the decision boundary is likely to be nonlinear.
60
+
Because the goal of this experiment is to predict automobile prices, and because the label column (price) contains real numbers, a regression model is a good choice. Considering that the number of features is relatively small (less than 100) and these features aren't sparse, the decision boundary is likely to be nonlinear.
61
61
62
-
To compare the performance of different algorithms, we use two nonlinear algorithms, **Boosted Decision Tree Regression** and **Decision Forest Regression**, to build models. Both algorithms have parameters that you can change, but this sample uses the default values for this pipeline.
62
+
To compare the performance of different algorithms, we use two nonlinear algorithms, **Boosted Decision Tree Regression** and **Decision Forest Regression**, to build models. Both algorithms have parameters that you can change, but this sample uses the default values for this experiment.
63
63
64
64
Use the **Split Data** module to randomly divide the input data so that the training dataset contains 70% of the original data and the testing dataset contains 30% of the original data.
65
65
@@ -69,7 +69,7 @@ You use two different sets of randomly chosen data to train and then test the mo
69
69
70
70
After the model is trained, use the **Score Model** and **Evaluate Model** modules to generate predicted results and evaluate the models. **Score Model** generates predictions for the test dataset by using the trained model. Then pass the scores to **Evaluate Model** to generate evaluation metrics.
71
71
72
-
In this pipeline, you use two instances of **Evaluate Model** to compare two pairs of models.
72
+
In this experiment, you use two instances of **Evaluate Model** to compare two pairs of models.
73
73
74
74
First, compare two algorithms on the training dataset.
75
75
Second, compare two algorithms on the testing dataset.
0 commit comments