You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
#customer intent: As a data scientist or engineer, I want to optimize my cost for training learning modules.
14
14
---
15
15
@@ -51,7 +51,7 @@ Because these compute pools are inside of Azure's IaaS infrastructure, you can d
51
51
52
52
Autoscaling clusters based on the requirements of your workload helps reduce your costs so you only use what you need.
53
53
54
-
AmlCompute clusters are designed to scale dynamically based on your workload. The cluster can be scaled up to the maximum number of nodes you configure. As each job completes, the cluster releases nodes and scale to your configured minimum node count.
54
+
AmlCompute clusters are designed to scale dynamically based on your workload. The cluster can be scaled up to the maximum number of nodes you configure. As each job finishes, the cluster releases nodes and scales to your configured minimum node count.
@@ -60,7 +60,7 @@ You can also configure the amount of time the node is idle before scale down. By
60
60
- If you perform less iterative experimentation, reduce this time to save costs.
61
61
- If you perform highly iterative dev/test experimentation, you might need to increase the time so that you don't pay for constant scaling up and down after each change to your training script or environment.
62
62
63
-
You can configure AmlCompute clusters your changing workload requirements by using:
63
+
You can configure AmlCompute clusters for your changing workload requirements by using:
64
64
65
65
- The Azure portal
66
66
- The [AmlCompute SDK class](/python/api/azure-ai-ml/azure.ai.ml.entities.amlcompute)
@@ -81,14 +81,14 @@ For more information, see [Autoscale online endpoints](how-to-autoscale-endpoint
81
81
82
82
## Set quotas on resources
83
83
84
-
AmlCompute comes with a quota, or limit, configuration. This quota is by VM family, for example, Dv2 series, NCv3 series. The quota varies by region for each subscription. Subscriptions start with small defaults. Use this setting to control the amount of Amlcompute resources available to be spun up in your subscription. For more information, see [Azure Machine Learning Compute](how-to-manage-quotas.md#azure-machine-learning-compute).
84
+
AmlCompute comes with a quota, or limit, configuration. This quota is by VM family, for example, Dv2 series or NCv3 series. The quota varies by region for each subscription. Subscriptions start with small defaults. Use this setting to control the amount of AmlCompute resources available to be spun up in your subscription. For more information, see [Azure Machine Learning Compute](how-to-manage-quotas.md#azure-machine-learning-compute).
85
85
86
86
Also, you can configure workspace level quota by VM family for each workspace within a subscription. This approach gives you more granular control on the costs that each workspace might incur and restricts certain VM families. For more information, see [Workspace-level quotas](how-to-manage-quotas.md#workspace-level-quotas).
87
87
88
88
To set quotas at the workspace level:
89
89
90
90
1. Open the [Azure portal](https://portal.azure.com) and then select any workspace in your subscription.
91
-
1. Select **Support + troubleshooting** > **Usages + quotas** in the workspace menu.
91
+
1. Select **Support + Troubleshooting** > **Usage + quotas** in the workspace menu.
92
92
1. Select **View quota** to view quotas in Azure Machine Learning studio.
93
93
1. From this page, you can find your subscription and region in order to set quotas.
94
94
@@ -102,7 +102,7 @@ Here are a few options that you have:
102
102
103
103
- Define a parameter called `max_run_duration_seconds` in your RunConfiguration to control the maximum duration a run can extend to on the compute you choose, either local or remote cloud compute.
104
104
- For *hyperparameter tuning*, define an early termination policy from a Bandit policy, a Median stopping policy, or a Truncation selection policy. To further control hyperparameter sweeps, use parameters such as `max_total_runs` or `max_duration_minutes`. For more information, see [Specify early termination policy](how-to-tune-hyperparameters.md#early-termination).
105
-
- For automated machine learning, set similar termination policies using the `enable_early_stopping` flag. Also use properties such as `iteration_timeout_minutes` and `experiment_timeout_minutes` to control the maximum duration of a job or for the entire experiment. For more information, see [Exit criteria](how-to-configure-auto-train.md#exit-criteria).
105
+
- For automated machine learning, set similar termination policies using the `enable_early_stopping` flag. You can also use properties such as `iteration_timeout_minutes` and `experiment_timeout_minutes` to control the maximum duration of a job or for the entire experiment. For more information, see [Exit criteria](how-to-configure-auto-train.md#exit-criteria).
@@ -141,7 +141,7 @@ For hybrid cloud scenarios like those that use Azure ExpressRoute, it can someti
141
141
142
142
## Delete failed deployments
143
143
144
-
Managed online endpoint uses VMs for the deployments. If you submitted request to create an online deployment and it failed, the request might have passed the stage when compute is created. In that case, the failed deployment would incur charges. When you finish debugging or investigation for the failure, delete the failed deployments to save the cost.
144
+
Managed online endpoints use VMs for the deployments. If you submitted request to create an online deployment and it failed, the request might have passed the stage when compute is created. In that case, the failed deployment would incur charges. When you finish debugging or investigation for the failure, delete the failed deployments to save the cost.
0 commit comments