You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
#Customer intent: As a professional data scientist, I want to know how to build and deploy a model with Azure Machine Learning by using Python in a Jupyter Notebook.
This tutorial series shows how features seamlessly integrate all phases of the machine learning lifecycle: prototyping, training, and operationalization.
21
21
22
-
The first tutorial showed how to create a feature set specification with custom transformations, and then use that feature set to generate training data, enable materialization, and perform a backfill. The second tutorial showed how to enable materialization, and perform a backfill. It also showed how to experiment with features, as a way to improve model performance.
22
+
The first tutorial showed how to create a feature set specification with custom transformations. It then showed how to use that feature set to generate training data, enable materialization, and perform a backfill. The second tutorial showed how to enable materialization and perform a backfill. It also showed how to experiment with features, as a way to improve model performance.
23
23
24
24
This tutorial explains how to:
25
25
@@ -35,27 +35,27 @@ Before you proceed with this tutorial, be sure to complete the first and second
35
35
36
36
1. Configure the Azure Machine Learning Spark notebook.
37
37
38
-
To run this tutorial, you can create a new notebook and execute the instructions step by step. You can also open and run the existing notebook named *3. Enable recurrent materialization and run batch inference*. You can find that notebook, and all the notebooks in this series, in the *featurestore_sample/notebooks* directory. You can choose *sdk_only* or *sdk_and_cli*. Keep this tutorial open and refer to it for documentation links and more explanation.
38
+
To run this tutorial, you can create a new notebook and execute the instructions, step by step. You can also open and run the existing notebook named *3. Enable recurrent materialization and run batch inference*. You can find that notebook, and all the notebooks in this series, in the *featurestore_sample/notebooks* directory. You can choose *sdk_only* or *sdk_and_cli*. Keep this tutorial open and refer to it for documentation links and more explanation.
39
39
40
40
1. In the **Compute** dropdown list in the top nav, select **Serverless Spark Compute** under **Azure Machine Learning Serverless Spark**.
41
41
42
-
2. Configure the session:
42
+
1. Configure the session:
43
43
44
44
1. Select **Configure session** in the top status bar.
45
-
2. Select the **Python packages** tab.
46
-
3. Select **Upload conda file**.
47
-
4. Select the `azureml-examples/sdk/python/featurestore-sample/project/env/online.yml` file from your local machine.
48
-
5. Optionally, increase the session time-out (idle time) to avoid frequent prerequisite reruns.
45
+
1. Select the **Python packages** tab.
46
+
1. Select **Upload conda file**.
47
+
1. Select the `azureml-examples/sdk/python/featurestore-sample/project/env/online.yml` file from your local machine.
48
+
1. Optionally, increase the session time-out (idle time) to avoid frequent prerequisite reruns.
1. Paste the `Data` field value in the following cell, with separate name and version values. The last character is the version, preceded by a colon (`:`).
150
150
1. Note the `predict_is_fraud` column that the batch inference pipeline generated.
151
151
152
-
In the batch inference pipeline (*/project/fraud_mode/pipelines/batch_inference_pipeline.yaml*) outputs, because you didn't provide `name` or `version` values for `outputs` of `inference_step`, the system created an untracked data asset with a GUID as the name value and `1` as the version value. In this cell, you derive and then display the data path from the asset.
152
+
In the batch inference pipeline (*/project/fraud_mode/pipelines/batch_inference_pipeline.yaml*) outputs, the system created an untracked data asset with a GUID as the name value and `1` as the version value. This happened because you didn't provide `name` or `version` values for `outputs` of `inference_step`. In this cell, you derive and then display the data path from the asset.
0 commit comments