Skip to content

Commit 76bb104

Browse files
authored
Merge pull request #76548 from j-martens/patch-448
Update how-to-deploy-inferencing-gpus.md
2 parents d0264f5 + 6b610c6 commit 76bb104

File tree

1 file changed

+7
-10
lines changed

1 file changed

+7
-10
lines changed

articles/machine-learning/service/how-to-deploy-inferencing-gpus.md

Lines changed: 7 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
---
2-
title: How to deploy a deep learning model for inferencing with GPU
2+
title: Deploy model for inferencing with GPU
33
titleSuffix: Azure Machine Learning service
44
description: Learn how to deploy a deep learning model as a web service that uses a GPU for inferencing. In this article, a Tensorflow model is deployed to an Azure Kubernetes Service cluster. The cluster uses a GPU-enabled VM to host the web service and score inferencing requests.
55
services: machine-learning
@@ -12,26 +12,23 @@ ms.reviewer: larryfr
1212
ms.date: 05/02/2019
1313
---
1414

15-
# How to do GPU inferencing
15+
# Deploy a deep learning model for inferencing with GPU
1616

1717
Learn how to use GPU inferencing for a machine learning model deployed as a web service. In this article, you learn how to use the Azure Machine Learning service to deploy an example Tensorflow deep learning model. The model is deployed to an Azure Kubernetes Service (AKS) cluster that uses a GPU-enabled VM to host the service. When requests are sent to the service, the model uses the GPU to perform inferencing.
1818

1919
GPUs offer performance advantages over CPUs on highly parallelizable computation. Training and inferencing deep learning models (especially for large batches of requests) are excellent use cases for GPUs.
2020

21-
This example will show you how to deploy a TensorFlow saved model to Azure Machine Learning.
21+
This example will show you how to deploy a TensorFlow saved model to Azure Machine Learning by:
22+
* Creating a GPU-enabled AKS cluster
23+
* Deploying a model with Tensorflow-GPU
2224

23-
## Goals and prerequisites
25+
## Prerequisites
2426

25-
Follow the instructions to:
26-
* Create a GPU enabled AKS cluster
27-
* Deploy a model with Tensorflow-GPU
28-
29-
Prerequisites:
3027
* Azure Machine Learning services workspace
3128
* Python
3229
* Tensorflow SavedModel registered. To learn how to register models see [Deploy Models](https://docs.microsoft.com/azure/machine-learning/service/how-to-deploy-and-where#registermodel)
3330

34-
This article is based on [Deploying Tensorflow Models to AKS](https://github.com/Azure/MachineLearningNotebooks/blob/master/how-to-use-azureml/deployment/production-deploy-to-aks-gpu/production-deploy-to-aks-gpu.ipynb), which uses TensorFlow saved models and deploys to an AKS cluster. However, with small changes to the scoring file and environment file it is applicable to any machine learning framework which support GPUs.
31+
This article is based on Jupyter notebook, [Deploying Tensorflow Models to AKS](https://github.com/Azure/MachineLearningNotebooks/blob/master/how-to-use-azureml/deployment/production-deploy-to-aks-gpu/production-deploy-to-aks-gpu.ipynb), which uses TensorFlow saved models and deploys to an AKS cluster. However, with small changes to the scoring file and environment file it is applicable to any machine learning framework which support GPUs.
3532

3633
## Provision AKS cluster with GPUs
3734
Azure has many different GPU options, all of which can be used for Inferencing. See [the list of N Series](https://azure.microsoft.com/pricing/details/virtual-machines/linux/#n-series) for a full breakdown of capabilities and costs.

0 commit comments

Comments
 (0)