Skip to content

Commit 66bec46

Browse files
authored
Merge pull request #106856 from nibaccam/dsets-maintenance
Data & SKlearn | Freshness update
2 parents 8350100 + 81f761e commit 66bec46

File tree

3 files changed

+16
-13
lines changed

3 files changed

+16
-13
lines changed

articles/machine-learning/how-to-train-scikit-learn.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ ms.subservice: core
88
ms.topic: conceptual
99
ms.author: maxluk
1010
author: maxluk
11-
ms.date: 08/02/2019
11+
ms.date: 03/09/2020
1212
ms.custom: seodec18
1313

1414
#Customer intent: As a Python scikit-learn developer, I need to combine open-source with a cloud platform to train, evaluate, and deploy my machine learning models at scale.
@@ -37,7 +37,7 @@ Run this code on either of these environments:
3737
- [Create a workspace configuration file](how-to-configure-environment.md#workspace).
3838
- Download the dataset and sample script file
3939
- [iris dataset](https://archive.ics.uci.edu/ml/datasets/iris)
40-
- [`train_iris.py`](https://github.com/Azure/MachineLearningNotebooks/tree/master/how-to-use-azureml/ml-frameworks/scikit-learn/training/train-hyperparameter-tune-deploy-with-sklearn)
40+
- [train_iris.py](https://github.com/Azure/MachineLearningNotebooks/tree/master/how-to-use-azureml/ml-frameworks/scikit-learn/training/train-hyperparameter-tune-deploy-with-sklearn)
4141
- You can also find a completed [Jupyter Notebook version](https://github.com/Azure/MachineLearningNotebooks/blob/master/how-to-use-azureml/ml-frameworks/scikit-learn/training/train-hyperparameter-tune-deploy-with-sklearn/train-hyperparameter-tune-deploy-with-sklearn.ipynb) of this guide on the GitHub samples page. The notebook includes an expanded section covering intelligent hyperparameter tuning and retrieving the best model by primary metrics.
4242

4343
## Set up the experiment
@@ -155,7 +155,7 @@ run = experiment.submit(estimator)
155155
run.wait_for_completion(show_output=True)
156156
```
157157

158-
As the Run is executed, it goes through the following stages:
158+
As the run is executed, it goes through the following stages:
159159

160160
- **Preparing**: A docker image is created according to the TensorFlow estimator. The image is uploaded to the workspace's container registry and cached for later runs. Logs are also streamed to the run history and can be viewed to monitor progress.
161161

@@ -177,7 +177,7 @@ import joblib
177177
joblib.dump(svm_model_linear, 'model.joblib')
178178
```
179179

180-
Register the model to your workspace with the following code. By specifying the parameters `model_framework`, `model_framework_version`, and `resource_configuration`, no-code model deployment becomes available. This allows you to directly deploy your model as a web service from the registered model, and the `ResourceConfiguration` object defines the compute resource for the web service.
180+
Register the model to your workspace with the following code. By specifying the parameters `model_framework`, `model_framework_version`, and `resource_configuration`, no-code model deployment becomes available. This allows you to directly deploy your model as a web service from the registered model, and the [`ResourceConfiguration`](https://docs.microsoft.com/python/api/azureml-core/azureml.core.resource_configuration.resourceconfiguration?view=azure-ml-py) object defines the compute resource for the web service.
181181

182182
```Python
183183
from azureml.core import Model
@@ -197,7 +197,7 @@ contains a section on registering models, but you can skip directly to [creating
197197

198198
### (Preview) No-code model deployment
199199

200-
Instead of the traditional deployment route, you can also use the no-code deployment feature (preview)for scikit-learn. No-code model deployment is supported for all built-in scikit-learn model types. By registering your model as shown above with the `model_framework`, `model_framework_version`, and `resource_configuration` parameters, you can simply use the `deploy()` static function to deploy your model.
200+
Instead of the traditional deployment route, you can also use the no-code deployment feature (preview)for scikit-learn. No-code model deployment is supported for all built-in scikit-learn model types. By registering your model as shown above with the `model_framework`, `model_framework_version`, and `resource_configuration` parameters, you can simply use the [`deploy()`](https://docs.microsoft.com/python/api/azureml-core/azureml.core.model%28class%29?view=azure-ml-py#deploy-workspace--name--models--inference-config-none--deployment-config-none--deployment-target-none--overwrite-false-) static function to deploy your model.
201201

202202
```python
203203
web_service = Model.deploy(ws, "scikit-learn-service", [model])

articles/machine-learning/how-to-train-with-datasets.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -10,16 +10,16 @@ ms.author: sihhu
1010
author: MayMSFT
1111
manager: cgronlun
1212
ms.reviewer: nibaccam
13-
ms.date: 09/25/2019
13+
ms.date: 03/09/2020
1414

15-
# Customer intent: As an experienced Python developer, I need to make my data available to my remote compute to train my machine learning models.
15+
# Customer intent: As an experienced Python developer, I need to make my data available to my local or remote compute to train my machine learning models.
1616

1717
---
1818

1919
# Train with datasets in Azure Machine Learning
2020
[!INCLUDE [applies-to-skus](../../includes/aml-applies-to-basic-enterprise-sku.md)]
2121

22-
In this article, you learn the two ways to consume [Azure Machine Learning datasets](https://docs.microsoft.com/python/api/azureml-core/azureml.core.dataset%28class%29?view=azure-ml-py) in remote experiment training runs without worrying about connection strings or data paths.
22+
In this article, you learn the two ways to consume [Azure Machine Learning datasets](https://docs.microsoft.com/python/api/azureml-core/azureml.core.dataset%28class%29?view=azure-ml-py) in a remote experiment training runs without worrying about connection strings or data paths.
2323

2424
- Option 1: If you have structured data, create a TabularDataset and use it directly in your training script.
2525

@@ -31,7 +31,7 @@ Azure Machine Learning datasets provide a seamless integration with Azure Machin
3131

3232
To create and train with datasets, you need:
3333

34-
* An Azure subscription. If you dont have an Azure subscription, create a free account before you begin. Try the [free or paid version of Azure Machine Learning](https://aka.ms/AMLFree) today.
34+
* An Azure subscription. If you don't have an Azure subscription, create a free account before you begin. Try the [free or paid version of Azure Machine Learning](https://aka.ms/AMLFree) today.
3535

3636
* An [Azure Machine Learning workspace](how-to-manage-workspace.md).
3737

articles/machine-learning/how-to-version-track-datasets.md

Lines changed: 7 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ ms.topic: conceptual
99
ms.author: sihhu
1010
author: sihhu
1111
ms.reviewer: nibaccam
12-
ms.date: 11/04/2019
12+
ms.date: 03/09/2020
1313
ms.custom:
1414

1515
# Customer intent: As a data scientist, I want to version and track datasets so I can use and share them across multiple machine learning experiments.
@@ -57,6 +57,7 @@ titanic_ds = titanic_ds.register(workspace = workspace,
5757
description = 'titanic training data',
5858
create_new_version = True)
5959
```
60+
You can also register a new version of a dataset at
6061

6162
### Retrieve a dataset by name
6263

@@ -117,7 +118,7 @@ dataset2.register(workspace = workspace,
117118

118119
You can use a dataset as the input and output of each Machine Learning pipeline step. When you rerun pipelines, the output of each pipeline step is registered as a new dataset version.
119120

120-
Because Machine Learning pipelines populate the output of each step into a new folder every time the pipeline reruns, the versioned output datasets are reproducible.
121+
Because Machine Learning pipelines populate the output of each step into a new folder every time the pipeline reruns, the versioned output datasets are reproducible. Learn more about [datasets in pipelines](how-to-create-your-first-pipeline.md#steps).
121122

122123
```Python
123124
from azureml.core import Dataset
@@ -166,7 +167,7 @@ input_dataset = inputs[0]['dataset']
166167
input_dataset.to_path()
167168
```
168169

169-
You can also find the `input_datasets` from experiments by using [Azure Machine Learning studio](https://ml.azure.com/).
170+
You can also find the `input_datasets` from experiments by using https://ml.azure.com/.
170171

171172
The following image shows where to find the input dataset of an experiment on Azure Machine Learning studio. For this example, go to your **Experiments** pane and open the **Properties** tab for a specific run of your experiment, `keras-mnist`.
172173

@@ -180,7 +181,9 @@ model = run.register_model(model_name='keras-mlp-mnist',
180181
datasets =[('training data',train_dataset)])
181182
```
182183

183-
After registration, you can see the list of models registered with the dataset by using Python or [Azure Machine Learning studio](https://ml.azure.com/). The following view is from the **Datasets** pane under **Assets**. Select the dataset and then select the **Models** tab for a list of the models that are registered with the dataset.
184+
After registration, you can see the list of models registered with the dataset by using Python or go to https://ml.azure.com/.
185+
186+
The following view is from the **Datasets** pane under **Assets**. Select the dataset and then select the **Models** tab for a list of the models that are registered with the dataset.
184187

185188
![Input datasets models](./media/how-to-version-track-datasets/dataset-models.png)
186189

0 commit comments

Comments
 (0)