You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/machine-learning/service/how-to-deploy-inferencing-gpus.md
+7-10Lines changed: 7 additions & 10 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,5 +1,5 @@
1
1
---
2
-
title: How to deploy a deep learning model for inferencing with GPU
2
+
title: Deploy model for inferencing with GPU
3
3
titleSuffix: Azure Machine Learning service
4
4
description: Learn how to deploy a deep learning model as a web service that uses a GPU for inferencing. In this article, a Tensorflow model is deployed to an Azure Kubernetes Service cluster. The cluster uses a GPU-enabled VM to host the web service and score inferencing requests.
5
5
services: machine-learning
@@ -12,26 +12,23 @@ ms.reviewer: larryfr
12
12
ms.date: 05/02/2019
13
13
---
14
14
15
-
# How to do GPU inferencing
15
+
# Deploy a deep learning model for inferencing with GPU
16
16
17
17
Learn how to use GPU inferencing for a machine learning model deployed as a web service. In this article, you learn how to use the Azure Machine Learning service to deploy an example Tensorflow deep learning model. The model is deployed to an Azure Kubernetes Service (AKS) cluster that uses a GPU-enabled VM to host the service. When requests are sent to the service, the model uses the GPU to perform inferencing.
18
18
19
19
GPUs offer performance advantages over CPUs on highly parallelizable computation. Training and inferencing deep learning models (especially for large batches of requests) are excellent use cases for GPUs.
20
20
21
-
This example will show you how to deploy a TensorFlow saved model to Azure Machine Learning.
21
+
This example will show you how to deploy a TensorFlow saved model to Azure Machine Learning by:
22
+
* Creating a GPU-enabled AKS cluster
23
+
* Deploying a model with Tensorflow-GPU
22
24
23
-
## Goals and prerequisites
25
+
## Prerequisites
24
26
25
-
Follow the instructions to:
26
-
* Create a GPU enabled AKS cluster
27
-
* Deploy a model with Tensorflow-GPU
28
-
29
-
Prerequisites:
30
27
* Azure Machine Learning services workspace
31
28
* Python
32
29
* Tensorflow SavedModel registered. To learn how to register models see [Deploy Models](https://docs.microsoft.com/azure/machine-learning/service/how-to-deploy-and-where#registermodel)
33
30
34
-
This article is based on [Deploying Tensorflow Models to AKS](https://github.com/Azure/MachineLearningNotebooks/blob/master/how-to-use-azureml/deployment/production-deploy-to-aks-gpu/production-deploy-to-aks-gpu.ipynb), which uses TensorFlow saved models and deploys to an AKS cluster. However, with small changes to the scoring file and environment file it is applicable to any machine learning framework which support GPUs.
31
+
This article is based on Jupyter notebook, [Deploying Tensorflow Models to AKS](https://github.com/Azure/MachineLearningNotebooks/blob/master/how-to-use-azureml/deployment/production-deploy-to-aks-gpu/production-deploy-to-aks-gpu.ipynb), which uses TensorFlow saved models and deploys to an AKS cluster. However, with small changes to the scoring file and environment file it is applicable to any machine learning framework which support GPUs.
35
32
36
33
## Provision AKS cluster with GPUs
37
34
Azure has many different GPU options, all of which can be used for Inferencing. See [the list of N Series](https://azure.microsoft.com/pricing/details/virtual-machines/linux/#n-series) for a full breakdown of capabilities and costs.
0 commit comments