You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/aks/gpu-cluster.md
+6-6Lines changed: 6 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -43,7 +43,7 @@ There are two options for adding the NVIDIA device plugin:
43
43
44
44
### Update your cluster to use the AKS GPU image (preview)
45
45
46
-
AKS provides is providing a fully configured AKS image that already contains the [NVIDIA device plugin for Kubernetes][nvidia-github].
46
+
AKS provides a fully configured AKS image that already contains the [NVIDIA device plugin for Kubernetes][nvidia-github].
47
47
48
48
Register the `GPUDedicatedVHDPreview` feature:
49
49
@@ -93,7 +93,7 @@ az aks nodepool add \
93
93
--max-count 3
94
94
```
95
95
96
-
The above command adds a node pool named *gpunp* to the *myAKSCluster* in the *myResourceGroup* resource group. The command also sets the VM size for the nodes in the node pool to *Standard_NC6*, enables the cluster autoscaler, configures the cluster autoscaler to maintain a minimum of one node and a maximum of three nodes in the node pool, specifies a specialized AKS GPU image nodes on your new node pool, and specifies a *sku=gpu:NoSchedule* taint for the node pool.
96
+
The above command adds a node pool named *gpunp* to the *myAKSCluster* in the *myResourceGroup* resource group. The command also sets the VM size for the node in the node pool to *Standard_NC6*, enables the cluster autoscaler, configures the cluster autoscaler to maintain a minimum of one node and a maximum of three nodes in the node pool, specifies a specialized AKS GPU image nodes on your new node pool, and specifies a *sku=gpu:NoSchedule* taint for the node pool.
97
97
98
98
> [!NOTE]
99
99
> A taint and VM size can only be set for node pools during node pool creation, but the autoscaler settings can be updated at any time.
@@ -408,8 +408,8 @@ For more information about running machine learning (ML) workloads on Kubernetes
408
408
409
409
For information on using Azure Kubernetes Service with Azure Machine Learning, see the following articles:
410
410
411
-
*[Deploy a model to Azure Kubernetes Service][azureml-aks].
412
-
*[Deploy a deep learning model for inference with GPU][azureml-gpu].
411
+
*[Configure a Kubernetes cluster for ML model training or deployment][azureml-aks].
412
+
*[Deploy a model with an online endpoint][azureml-deploy].
413
413
*[High-performance serving with Triton Inference Server][azureml-triton].
414
414
415
415
<!-- LINKS - external -->
@@ -434,7 +434,7 @@ For information on using Azure Kubernetes Service with Azure Machine Learning, s
Copy file name to clipboardExpand all lines: articles/machine-learning/reference-machine-learning-cloud-parity.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -23,7 +23,7 @@ In the list of global Azure regions, there are several regions that serve specif
23
23
* Azure Government regions **US-Arizona** and **US-Virginia**.
24
24
* Azure China 21Vianet region **China-East-2**.
25
25
26
-
Azure Machine Learning is still in development in Airgap Regions.
26
+
Azure Machine Learning is still in development in air-gap Regions.
27
27
28
28
The information in the rest of this document provides information on what features of Azure Machine Learning are available in these regions, along with region-specific information on using these features.
29
29
## Azure Government
@@ -76,7 +76,7 @@ The information in the rest of this document provides information on what featur
76
76
|**Machine learning lifecycle**||||
77
77
|[Model profiling](v1/how-to-deploy-profile-model.md)| GA | YES | PARTIAL |
78
78
|[The Azure ML CLI 1.0](v1/reference-azure-machine-learning-cli.md)| GA | YES | YES |
79
-
|[FPGA-based Hardware Accelerated Models](how-to-deploy-fpga-web-service.md)| GA | NO | NO |
79
+
|[FPGA-based Hardware Accelerated Models](./v1/how-to-deploy-fpga-web-service.md)| GA | NO | NO |
80
80
|[Visual Studio Code integration](how-to-setup-vs-code.md)| Public Preview | NO | NO |
81
81
|[Event Grid integration](how-to-use-event-grid.md)| Public Preview | NO | NO |
82
82
|[Integrate Azure Stream Analytics with Azure Machine Learning](../stream-analytics/machine-learning-udf.md)| Public Preview | NO | NO |
@@ -99,7 +99,7 @@ The information in the rest of this document provides information on what featur
99
99
|**Inference**||||
100
100
| Managed online endpoints | GA | YES | YES |
101
101
|[Batch inferencing](tutorial-pipeline-batch-scoring-classification.md)| GA | YES | YES |
102
-
|[Azure Stack Edge with FPGA](how-to-deploy-fpga-web-service.md#deploy-to-a-local-edge-server)| Public Preview | NO | NO |
102
+
|[Azure Stack Edge with FPGA](./v1/how-to-deploy-fpga-web-service.md#deploy-to-a-local-edge-server)| Public Preview | NO | NO |
103
103
|**Other**||||
104
104
|[Open Datasets](../open-datasets/samples.md)| Public Preview | YES | YES |
105
105
|[Custom Cognitive Search](how-to-deploy-model-cognitive-search.md)| Public Preview | YES | YES |
In this article, you learn about FPGAs and how to deploy your ML models to an Azure FPGA using the [hardware-accelerated models Python package](/python/api/azureml-accel-models/azureml.accel) from [Azure Machine Learning](overview-what-is-azure-machine-learning.md).
20
+
In this article, you learn about FPGAs and how to deploy your ML models to an Azure FPGA using the [hardware-accelerated models Python package](/python/api/azureml-accel-models/azureml.accel) from [Azure Machine Learning](../overview-what-is-azure-machine-learning.md).
21
21
22
22
## What are FPGAs?
23
23
@@ -32,7 +32,7 @@ You can reconfigure FPGAs for different types of machine learning models. This f
32
32
|Processor| Abbreviation |Description|
33
33
|---|:-------:|------|
34
34
|Application-specific integrated circuits|ASICs|Custom circuits, such as Google's Tensor Processor Units (TPU), provide the highest efficiency. They can't be reconfigured as your needs change.|
35
-
|Field-programmable gate arrays|FPGAs|FPGAs, such as those available on Azure, provide performance close to ASICs. They are also flexible and reconfigurable over time, to implement new logic.|
35
+
|Field-programmable gate arrays|FPGAs|FPGAs, such as those available on Azure, provide performance close to ASICs. They're also flexible and reconfigurable over time, to implement new logic.|
36
36
|Graphics processing units|GPUs|A popular choice for AI computations. GPUs offer parallel processing capabilities, making it faster at image rendering than CPUs.|
37
37
|Central processing units|CPUs|General-purpose processors, the performance of which isn't ideal for graphics and video processing.|
38
38
@@ -50,7 +50,7 @@ Azure FPGAs are integrated with Azure Machine Learning. Azure can parallelize pr
50
50
51
51
To optimize latency and throughput, your client sending data to the FPGA model should be in one of the regions above (the one you deployed the model to).
52
52
53
-
The **PBS Family of Azure VMs** contains Intel Arria 10 FPGAs. It will show as "Standard PBS Family vCPUs" when you check your Azure quota allocation. The PB6 VM has six vCPUs and one FPGA. PB6 VM is automatically provisioned by Azure Machine Learning during model deployment to an FPGA. It is only used with Azure ML, and it cannot run arbitrary bitstreams. For example, you will not be able to flash the FPGA with bitstreams to do encryption, encoding, etc.
53
+
The **PBS Family of Azure VMs** contains Intel Arria 10 FPGAs. It will show as "Standard PBS Family vCPUs" when you check your Azure quota allocation. The PB6 VM has six vCPUs and one FPGA. PB6 VM is automatically provisioned by Azure Machine Learning during model deployment to an FPGA. It's only used with Azure ML, and it can't run arbitrary bitstreams. For example, you won't be able to flash the FPGA with bitstreams to do encryption, encoding, etc.
54
54
55
55
## Deploy models on FPGAs
56
56
@@ -60,9 +60,9 @@ In this example, you create a TensorFlow graph to preprocess the input image, ma
60
60
61
61
### Prerequisites
62
62
63
-
- An Azure subscription. If you do not have one, create a [pay-as-you-go](https://azure.microsoft.com/pricing/purchase-options/pay-as-you-go) account (free Azure accounts are not eligible for FPGA quota).
63
+
- An Azure subscription. If you don't have one, create a [pay-as-you-go](https://azure.microsoft.com/pricing/purchase-options/pay-as-you-go) account (free Azure accounts aren't eligible for FPGA quota).
64
64
65
-
- An Azure Machine Learning workspace and the Azure Machine Learning SDK for Python installed, as described in [Create a workspace](how-to-manage-workspace.md).
65
+
- An Azure Machine Learning workspace and the Azure Machine Learning SDK for Python installed, as described in [Create a workspace](../how-to-manage-workspace.md).
66
66
67
67
- The hardware-accelerated models package: `pip install --upgrade azureml-accel-models[cpu]`
68
68
@@ -151,7 +151,7 @@ Begin by using the [Azure Machine Learning SDK for Python](/python/api/overview/
151
151
print(output_tensors)
152
152
```
153
153
154
-
The following models are available listed with their classifier output tensors for inference if you used the default classifier.
154
+
The following models are listed with their classifier output tensors for inference if you used the default classifier.
155
155
156
156
+ Resnet50, QuantizedResnet50
157
157
```python
@@ -178,7 +178,7 @@ Begin by using the [Azure Machine Learning SDK for Python](/python/api/overview/
178
178
179
179
Before you can deploy to FPGAs, convert the model to the [ONNX](https://onnx.ai/) format.
180
180
181
-
1. [Register](concept-model-management-and-deployment.md) the model by using the SDKwith the ZIPfilein Azure Blob storage. Adding tags and other metadata about the model helps you keep track of your trained models.
181
+
1. [Register](../concept-model-management-and-deployment.md) the model by using the SDKwith the ZIPfilein Azure Blob storage. Adding tags and other metadata about the model helps you keep track of your trained models.
182
182
183
183
```python
184
184
from azureml.core.model import Model
@@ -221,7 +221,7 @@ Before you can deploy to FPGAs, convert the model to the [ONNX](https://onnx.ai/
221
221
222
222
### Containerize and deploy the model
223
223
224
-
Next, create a Docker image from the converted model and all dependencies. This Docker image can then be deployed and instantiated. Supported deployment targets include Azure Kubernetes Service (AKS) in the cloud or an edge device such as [Azure Data Box Edge](../databox-online/azure-stack-edge-overview.md). You can also add tags and descriptions for your registered Docker image.
224
+
Next, create a Docker image from the converted model and all dependencies. This Docker image can then be deployed and instantiated. Supported deployment targets include Azure Kubernetes Service (AKS) in the cloud or an edge device such as [Azure Azure Stack Edge](../../databox-online/azure-stack-edge-overview.md). You can also add tags and descriptions for your registered Docker image.
225
225
226
226
```python
227
227
from azureml.core.image import Image
@@ -295,7 +295,7 @@ Next, create a Docker image from the converted model and all dependencies. This
295
295
296
296
#### Deploy to a local edge server
297
297
298
-
All [Azure Data Box Edge devices](../databox-online/azure-stack-edge-overview.md) contain an FPGAfor running the model. Only one model can be running on the FPGA at one time. To run a different model, just deploy a new container. Instructions and sample code can be found in [this Azure Sample](https://github.com/Azure-Samples/aml-hardware-accelerated-models).
298
+
All [Azure Azure Stack Edge devices](../../databox-online/azure-stack-edge-overview.md) contain an FPGAfor running the model. Only one model can be running on the FPGA at one time. To run a different model, just deploy a new container. Instructions and sample code can be found in [this Azure Sample](https://github.com/Azure-Samples/aml-hardware-accelerated-models).
299
299
300
300
### Consume the deployed model
301
301
@@ -358,7 +358,7 @@ converted_model.delete()
358
358
359
359
## Next steps
360
360
361
-
+ Learn how to [secure your web services](./v1/how-to-secure-web-service.md) document.
361
+
+ Learn how to [secure your web services](how-to-secure-web-service.md) document.
362
362
363
363
+ Learn about FPGAand [Azure Machine Learning pricing and costs](https://azure.microsoft.com/pricing/details/machine-learning/).
0 commit comments