Skip to content

Commit 2874ccd

Browse files
authored
Merge pull request #245598 from ssalgadodev/patch-26
Update classification.md
2 parents b751bfa + fd875f7 commit 2874ccd

File tree

8 files changed

+100
-71
lines changed

8 files changed

+100
-71
lines changed

articles/machine-learning/component-reference-v2/classification.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ ms.topic: reference
99
ms.author: rasavage
1010
author: rsavage2
1111
ms.reviewer: ssalgadodev
12-
ms.date: 12/1/2022
12+
ms.date: 07/1/2023
1313
---
1414

1515
# AutoML Classification
@@ -34,35 +34,35 @@ AutoML creates a number of pipelines in parallel that try different algorithms a
3434

3535
1. For **classification**, you can also enable deep learning.
3636

37-
If deep learning is enabled, validation is limited to _train_validation split_. [Learn more about validation options](../v1/how-to-configure-cross-validation-data-splits.md).
37+
If deep learning is enabled, validation is limited to _train_validation split_.
3838

39-
40-
1. (Optional) View addition configuration settings: additional settings you can use to better control the training job. Otherwise, defaults are applied based on experiment selection and data.
39+
4. (Optional) View addition configuration settings: additional settings you can use to better control the training job. Otherwise, defaults are applied based on experiment selection and data.
4140

4241
Additional configurations|Description
4342
------|------
4443
Primary metric| Main metric used for scoring your model. [Learn more about model metrics](../how-to-configure-auto-train.md#primary-metric).
45-
Explain best model | Select to enable or disable, in order to show explanations for the recommended best model. <br> This functionality is not currently available for [certain forecasting algorithms](../v1/how-to-machine-learning-interpretability-automl.md#interpretability-during-training-for-the-best-model).
44+
Debug model via the Responsible AI dashboard | Generate a Responsible AI dashboard to do a holistic assessment and debugging of the recommended best model. This includes insights such as model explanations, fairness and performance explorer, data explorer, and model error analysis. [Learn more about how you can generate a Responsible AI dashboard.](../how-to-responsible-ai-insights-ui.md)
4645
Blocked algorithm| Select algorithms you want to exclude from the training job. <br><br> Allowing algorithms is only available for [SDK experiments](../how-to-configure-auto-train.md#supported-algorithms). <br> See the [supported algorithms for each task type](/python/api/azureml-automl-core/azureml.automl.core.shared.constants.supportedmodels).
4746
Exit criterion| When any of these criteria are met, the training job is stopped. <br> *Training job time (hours)*: How long to allow the training job to run. <br> *Metric score threshold*: Minimum metric score for all pipelines. This ensures that if you have a defined target metric you want to reach, you do not spend more time on the training job than necessary.
4847
Concurrency| *Max concurrent iterations*: Maximum number of pipelines (iterations) to test in the training job. The job will not run more than the specified number of iterations. Learn more about how automated ML performs [multiple child jobs on clusters](../how-to-configure-auto-train.md#multiple-child-runs-on-clusters).
4948

5049

5150
1. The **[Optional] Validate and test** form allows you to do the following.
5251

53-
1. Specify the type of validation to be used for your training job. [Learn more about cross validation](../v1/how-to-configure-cross-validation-data-splits.md#prerequisites).
52+
1. Specify the type of validation to be used for your training job.
5453

5554
1. Provide a test dataset (preview) to evaluate the recommended model that automated ML generates for you at the end of your experiment. When you provide test data, a test job is automatically triggered at the end of your experiment. This test job is only job on the best model that was recommended by automated ML.
5655

5756
>[!IMPORTANT]
5857
> Providing a test dataset to evaluate generated models is a preview feature. This capability is an [experimental](/python/api/overview/azure/ml/#stable-vs-experimental) preview feature, and may change at any time.
5958
6059
* Test data is considered a separate from training and validation, so as to not bias the results of the test job of the recommended model. [Learn more about bias during model validation](../concept-automated-ml.md#training-validation-and-test-data).
61-
* You can either provide your own test dataset or opt to use a percentage of your training dataset. Test data must be in the form of an [Azure Machine Learning TabularDataset](../v1/how-to-create-register-datasets.md#tabulardataset).
60+
* You can either provide your own test dataset or opt to use a percentage of your training dataset. Test data must be in the form of an [Azure Machine Learning TabularDataset](../how-to-create-data-assets.md).
6261
* The schema of the test dataset should match the training dataset. The target column is optional, but if no target column is indicated no test metrics are calculated.
6362
* The test dataset should not be the same as the training dataset or the validation dataset.
6463

6564

65+
6666
## Next steps
6767

6868
See the [set of components available](../component-reference/component-reference.md) to Azure Machine Learning.

articles/machine-learning/component-reference-v2/regression.md

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ ms.topic: reference
99
ms.author: rasavage
1010
author: rsavage2
1111
ms.reviewer: ssalgadodev
12-
ms.date: 12/1/2022
12+
ms.date: 07/17/2023
1313
---
1414

1515
# AutoML Regression
@@ -37,7 +37,7 @@ AutoML creates a number of pipelines in parallel that try different algorithms a
3737
Additional configurations|Description
3838
------|------
3939
Primary metric| Main metric used for scoring your model. [Learn more about model metrics](..//how-to-configure-auto-train.md#primary-metric).
40-
Explain best model | Select to enable or disable, in order to show explanations for the recommended best model. <br> This functionality is not currently available for [certain forecasting algorithms](../v1/how-to-machine-learning-interpretability-automl.md#interpretability-during-training-for-the-best-model).
40+
Debug model via the [Responsible AI dashboard](..//concept-responsible-ai-dashboard.md) | Generate a Responsible AI dashboard to do a holistic assessment and debugging of the recommended best model. This includes insights such as model explanations, fairness and performance explorer, data explorer, and model error analysis. [Learn more about how you can generate a Responsible AI dashboard.](../how-to-responsible-ai-insights-ui.md)
4141
Blocked algorithm| Select algorithms you want to exclude from the training job. <br><br> Allowing algorithms is only available for [SDK experiments](../how-to-configure-auto-train.md#supported-algorithms). <br> See the [supported algorithms for each task type](/python/api/azureml-automl-core/azureml.automl.core.shared.constants.supportedmodels).
4242
Exit criterion| When any of these criteria are met, the training job is stopped. <br> *Training job time (hours)*: How long to allow the training job to run. <br> *Metric score threshold*: Minimum metric score for all pipelines. This ensures that if you have a defined target metric you want to reach, you do not spend more time on the training job than necessary.
4343
Concurrency| *Max concurrent iterations*: Maximum number of pipelines (iterations) to test in the training job. The job will not run more than the specified number of iterations. Learn more about how automated ML performs [multiple child jobs on clusters](../how-to-configure-auto-train.md#multiple-child-runs-on-clusters).
@@ -46,7 +46,7 @@ AutoML creates a number of pipelines in parallel that try different algorithms a
4646

4747
1. The **[Optional] Validate and test** form allows you to do the following.
4848

49-
1. Specify the type of validation to be used for your training job. [Learn more about cross validation](../v1/how-to-configure-cross-validation-data-splits.md#prerequisites).
49+
1. Specify the type of validation to be used for your training job.
5050

5151

5252
1. Provide a test dataset (preview) to evaluate the recommended model that automated ML generates for you at the end of your experiment. When you provide test data, a test job is automatically triggered at the end of your experiment. This test job is only job on the best model that was recommended by automated ML.
@@ -55,11 +55,12 @@ AutoML creates a number of pipelines in parallel that try different algorithms a
5555
> Providing a test dataset to evaluate generated models is a preview feature. This capability is an [experimental](/python/api/overview/azure/ml/#stable-vs-experimental) preview feature, and may change at any time.
5656
5757
* Test data is considered a separate from training and validation, so as to not bias the results of the test job of the recommended model. [Learn more about bias during model validation](../concept-automated-ml.md#training-validation-and-test-data).
58-
* You can either provide your own test dataset or opt to use a percentage of your training dataset. Test data must be in the form of an [Azure Machine Learning TabularDataset](../v1/how-to-create-register-datasets.md#tabulardataset).
58+
* You can either provide your own test dataset or opt to use a percentage of your training dataset. Test data must be in the form of an [Azure Machine Learning TabularDataset](../how-to-create-data-assets.md).
5959
* The schema of the test dataset should match the training dataset. The target column is optional, but if no target column is indicated no test metrics are calculated.
6060
* The test dataset should not be the same as the training dataset or the validation dataset.
6161
* Forecasting jobs do not support train/test split.
6262

63+
6364
6465
6566

articles/machine-learning/how-to-understand-automated-ml.md

Lines changed: 19 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -8,14 +8,14 @@ ms.author: magoswam
88
ms.reviewer: ssalgado
99
ms.service: machine-learning
1010
ms.subservice: automl
11-
ms.date: 06/7/2023
11+
ms.date: 07/20/2023
1212
ms.topic: how-to
1313
ms.custom: contperf-fy21q2, automl, event-tier1-build-2022
1414
---
1515

1616
# Evaluate automated machine learning experiment results
1717

18-
In this article, learn how to evaluate and compare models trained by your automated machine learning (automated ML) experiment. Over the course of an automated ML experiment, many jobs are created and each job creates a model. For each model, automated ML generates evaluation metrics and charts that help you measure the model's performance.
18+
In this article, learn how to evaluate and compare models trained by your automated machine learning (automated ML) experiment. Over the course of an automated ML experiment, many jobs are created and each job creates a model. For each model, automated ML generates evaluation metrics and charts that help you measure the model's performance. You can further generate a Responsible AI dashboard to do a holistic assessment and debugging of the recommended best model by default. This includes insights such as model explanations, fairness and performance explorer, data explorer, model error analysis. Learn more about how you can generate a [Responsible AI dashboard.](how-to-responsible-ai-insights-ui.md)
1919

2020
For example, automated ML generates the following charts based on experiment type.
2121

@@ -49,7 +49,7 @@ After your automated ML experiment completes, a history of the jobs can be found
4949
The following steps and video, show you how to view the run history and model evaluation metrics and charts in the studio:
5050

5151
1. [Sign into the studio](https://ml.azure.com/) and navigate to your workspace.
52-
1. In the left menu, select **Runs**.
52+
1. In the left menu, select **Jobs**.
5353
1. Select your experiment from the list of experiments.
5454
1. In the table at the bottom of the page, select an automated ML job.
5555
1. In the **Models** tab, select the **Algorithm name** for the model you want to evaluate.
@@ -88,12 +88,12 @@ weighted_accuracy|Weighted accuracy is accuracy where each sample is weighted by
8888

8989
### Binary vs. multiclass classification metrics
9090

91-
Automated ML automatically detects if the data is binary and also allows users to activate binary classification metrics even if the data is multiclass by specifying a `true` class. Multiclass classification metrics will be reported no matter if a dataset has two classes or more than two classes. Binary classification metrics will only be reported when the data is binary, or the users activate the option.
91+
Automated ML automatically detects if the data is binary and also allows users to activate binary classification metrics even if the data is multiclass by specifying a `true` class. Multiclass classification metrics is reported no matter if a dataset has two classes or more than two classes. Binary classification metrics is only reported when the data is binary, or the users activate the option.
9292

9393
> [!Note]
9494
> When a binary classification task is detected, we use `numpy.unique` to find the set of labels and the later label will be used as the `true` class. Since there is a sorting procedure in `numpy.unique`, the choice of `true` class will be stable.
9595
96-
Note that multiclass classification metrics are intended for multiclass classification. When applied to a binary dataset, these metrics won't treat any class as the `true` class, as you might expect. Metrics that are clearly meant for multiclass are suffixed with `micro`, `macro`, or `weighted`. Examples include `average_precision_score`, `f1_score`, `precision_score`, `recall_score`, and `AUC`. For example, instead of calculating recall as `tp / (tp + fn)`, the multiclass averaged recall (`micro`, `macro`, or `weighted`) averages over both classes of a binary classification dataset. This is equivalent to calculating the recall for the `true` class and the `false` class separately, and then taking the average of the two.
96+
Note, multiclass classification metrics are intended for multiclass classification. When applied to a binary dataset, these metrics don't treat any class as the `true` class, as you might expect. Metrics that are clearly meant for multiclass are suffixed with `micro`, `macro`, or `weighted`. Examples include `average_precision_score`, `f1_score`, `precision_score`, `recall_score`, and `AUC`. For example, instead of calculating recall as `tp / (tp + fn)`, the multiclass averaged recall (`micro`, `macro`, or `weighted`) averages over both classes of a binary classification dataset. This is equivalent to calculating the recall for the `true` class and the `false` class separately, and then taking the average of the two.
9797

9898
Besides, although automatic detection of binary classification is supported, it is still recommended to always specify the `true` class manually to make sure the binary classification metrics are calculated for the correct class.
9999

@@ -309,11 +309,22 @@ The mAP, precision and recall values are logged at an epoch-level for image obje
309309
310310
![Epoch-level charts for object detection](./media/how-to-understand-automated-ml/image-object-detection-map.png)
311311

312-
## Model explanations and feature importances
312+
## Responsible AI dashboard for best recommended AutoML model (preview)
313+
314+
The Azure Machine Learning Responsible AI dashboard provides a single interface to help you implement Responsible AI in practice effectively and efficiently. Responsible AI dashboard is only supported using tabular data and is only supported on classification and regression models. It brings together several mature Responsible AI tools in the areas of:
315+
316+
* Model performance and fairness assessment
317+
* Data exploration
318+
* Machine learning interpretability
319+
* Error analysis
313320

314-
While model evaluation metrics and charts are good for measuring the general quality of a model, inspecting which dataset features a model used to make its predictions is essential when practicing responsible AI. That's why automated ML provides a model explanations dashboard to measure and report the relative contributions of dataset features. See how to [view the explanations dashboard in the Azure Machine Learning studio](how-to-use-automated-ml-for-ml-models.md#model-explanations-preview).
321+
While model evaluation metrics and charts are good for measuring the general quality of a model, operations such as inspecting you model’s fairness, viewing its explanations (also known as which dataset features a model used to make its predictions), inspecting its errors (what are the blindspots of the model) are essential when practicing responsible AI. That's why automated ML provides a Responsible AI dashboard to help you observe a variety of insights for your model. See how to view the Responsible AI dashboard in the [Azure Machine Learning studio.](how-to-use-automated-ml-for-ml-models.md#responsible-ai-dashboard-preview)
322+
323+
See how you can generate this [dashboard via the UI or the SDK.](how-to-responsible-ai-insights-sdk-cli.md)
324+
325+
## Model explanations and feature importances
315326

316-
For a code first experience, see how to set up [model explanations for automated ML experiments with the Azure Machine Learning Python SDK (v1)](./v1/how-to-machine-learning-interpretability-automl.md).
327+
While model evaluation metrics and charts are good for measuring the general quality of a model, inspecting which dataset features a model used to make its predictions is essential when practicing responsible AI. That's why automated ML provides a model explanations dashboard to measure and report the relative contributions of dataset features. See how to [view the explanations dashboard in the Azure Machine Learning studio](how-to-use-automated-ml-for-ml-models.md#responsible-ai-dashboard-preview).
317328

318329
> [!NOTE]
319330
> Interpretability, best model explanation, is not available for automated ML forecasting experiments that recommend the following algorithms as the best model or ensemble:

0 commit comments

Comments
 (0)