You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: includes/aml-compute-target-deploy.md
+7-7Lines changed: 7 additions & 7 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -14,12 +14,12 @@ ms.date: 10/21/2021
14
14
15
15
The compute target you use to host your model will affect the cost and availability of your deployed endpoint. Use this table to choose an appropriate compute target.
16
16
17
-
| Compute target | Used for | GPU support |FPGA support |Description |
18
-
| ----- | ----- | ----- | ----- |----- |
19
-
| [Local web service](../articles/machine-learning/v1/how-to-deploy-local-container-notebook-vm.md) | Testing/debugging | | | Use for limited testing and troubleshooting. Hardware acceleration depends on use of libraries in the local system.
|[Azure Container Instances](../articles/machine-learning/v1/how-to-deploy-azure-container-instance.md)| Real-time inference <br/><br/> Recommended for dev/test purposes only.| | |Use for low-scale CPU-based workloads that require less than 48 GB of RAM. Doesn't require you to manage a cluster. <br/><br/> Supported in the designer. |
22
-
|[Azure Machine Learning compute clusters](../articles/machine-learning/tutorial-pipeline-batch-scoring-classification.md)| Batch inference |[Yes](../articles/machine-learning/tutorial-pipeline-batch-scoring-classification.md) (machine learning pipeline) | | Run batch scoring on serverless compute. Supports normal and low-priority VMs. No support for real-time inference.|
17
+
| Compute target | Used for | GPU support | Description |
18
+
| ----- | ----- | ----- | ----- |
19
+
| [Local web service](../articles/machine-learning/v1/how-to-deploy-local-container-notebook-vm.md) | Testing/debugging | | Use for limited testing and troubleshooting. Hardware acceleration depends on use of libraries in the local system.
20
+
|[Azure Machine Learning Kubernetes](../articles/machine-learning/how-to-attach-kubernetes-anywhere.md)| Real-time inference <br/><br/> Batch inference | Yes | Run inferencing workloads on on-premises, cloud, and edge Kubernetes clusters. |
21
+
|[Azure Container Instances](../articles/machine-learning/v1/how-to-deploy-azure-container-instance.md)| Real-time inference <br/><br/> Recommended for dev/test purposes only.| | Use for low-scale CPU-based workloads that require less than 48 GB of RAM. Doesn't require you to manage a cluster. <br/><br/> Supported in the designer. |
22
+
|[Azure Machine Learning compute clusters](../articles/machine-learning/tutorial-pipeline-batch-scoring-classification.md)| Batch inference |[Yes](../articles/machine-learning/tutorial-pipeline-batch-scoring-classification.md) (machine learning pipeline) | Run batch scoring on serverless compute. Supports normal and low-priority VMs. No support for real-time inference.|
23
23
24
24
> [!NOTE]
25
25
> Although compute targets like local, and Azure Machine Learning compute clusters support GPU for training and experimentation, using GPU for inference _when deployed as a web service_ is supported only on Azure Machine Learning Kubernetes.
@@ -29,4 +29,4 @@ The compute target you use to host your model will affect the cost and availabil
29
29
> When choosing a cluster SKU, first scale up and then scale out. Start with a machine that has 150% of the RAM your model requires, profile the result and find a machine that has the performance you need. Once you've learned that, increase the number of machines to fit your need for concurrent inference.
30
30
31
31
> [!NOTE]
32
-
> * Container instances are suitable only for small models less than 1 GB in size.
32
+
> * Container instances are suitable only for small models less than 1 GB in size.
0 commit comments