Skip to content

Commit 7dd282d

Browse files
authored
Fix typos and improve clarity in documentation
1 parent dfced3d commit 7dd282d

File tree

1 file changed

+11
-11
lines changed

1 file changed

+11
-11
lines changed

articles/ai-foundry/how-to/deploy-nvidia-inference-microservice.md

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ Get improved TCO (total cost of ownership) and performance with NVIDIA NIMs offe
2828

2929
- An [Azure AI Foundry project](create-projects.md).
3030

31-
- Ensure Marketplace purcharses are enabled for your azure subscription. Learn more about it [here](/azure/cost-management-billing/manage/enable-marketplace-purchases).
31+
- Ensure Marketplace purchases are enabled for your Azure subscription. Learn more about it [here](/azure/cost-management-billing/manage/enable-marketplace-purchases).
3232

3333
- Azure role-based access controls (Azure RBAC) are used to grant access to operations in Azure AI Foundry portal. To perform the steps in this article, your user account must be assigned a _custom role_ with the following permissions. User accounts assigned the _Owner_ or _Contributor_ role for the Azure subscription can also create NIM deployments. For more information on permissions, see [Role-based access control in Azure AI Foundry portal](../concepts/rbac-ai-foundry.md).
3434

@@ -50,7 +50,7 @@ Get improved TCO (total cost of ownership) and performance with NVIDIA NIMs offe
5050

5151
## NVIDIA NIM PayGo offer on Azure Marketplace by NVIDIA
5252

53-
NVIDIA NIMs available on Azure AI Foundry model catalog can be deployed with a suscription to the [NVIDIA NIM SaaS offer](https://aka.ms/nvidia-nims-plan) on Azure Marketplace. This offer includes a 90-day trial that applies to all NIMs associated with a particular SaaS subscription scoped to a Azure AI Foundry project, and has a PayGo price of $1 per GPU hour post the trial period.
53+
NVIDIA NIMs available on Azure AI Foundry model catalog can be deployed with a subscription to the [NVIDIA NIM SaaS offer](https://aka.ms/nvidia-nims-plan) on Azure Marketplace. This offer includes a 90-day trial that applies to all NIMs associated with a particular SaaS subscription scoped to an Azure AI Foundry project, and has a PayGo price of $1 per GPU hour post the trial period.
5454

5555
Azure AI Foundry enables a seamless purchase flow of the NVIDIA NIM offering on Marketplace from NVIDIA collection in the model catalog, and further deployment on managed compute.
5656

@@ -62,26 +62,26 @@ Get improved TCO (total cost of ownership) and performance with NVIDIA NIMs offe
6262

6363
:::image type="content" source="../media/how-to/deploy-nvidia-inference-microservice/nvidia-collections.png" alt-text="A screenshot showing how to filter by NVIDIA collections models in the catalog." lightbox="../media/how-to/deploy-nvidia-inference-microservice/nvidia-collections.png":::
6464

65-
1. Select the NVIDIA NIM of your choice. In this article, we will be using **Llama-3.3-70B-Instruct-NIM-microservice** as an example.
65+
1. Select the NVIDIA NIM of your choice. In this article, we are using **Llama-3.3-70B-Instruct-NIM-microservice** as an example.
6666
1. Select **Deploy**.
67-
1. Select one of the NVIDIA GPU based VM SKUs supported for the NIM, based on your intended workload. You will need to have quota in your Azure subscription.
68-
1. You can then customize your deployment configuration for the instance count, select an existing endpoint or create a new one, etc. For the example in this article, we will consider an instance count of **2** and create a new endpoint.
67+
1. Select one of the NVIDIA GPU based VM SKUs supported for the NIM, based on your intended workload. You need to have quota in your Azure subscription.
68+
1. You can then customize your deployment configuration for the instance count, select an existing endpoint or create a new one, etc. For the example in this article, we consider an instance count of **2** and create a new endpoint.
6969

7070
:::image type="content" source="../media/how-to/deploy-nvidia-inference-microservice/project-customization.png" alt-text="A screenshot showing project customization options in the deployment wizard." lightbox="../media/how-to/deploy-nvidia-inference-microservice/project-customization.png":::
7171

7272
1. Select **Next**
73-
1. You will then need to review the pricing breakdown for the NIM deployment, terms of use and license agreement associated with the NIM offer. The pricing breakdown will help inform what the aggregated pricing for the NIM software deployed would be, which is a function of number of NVIDIA GPUs in the VM instance selected in the previous steps. In addition to the applicable NIM software price, Azure Compute charges will also apply based on your deployment configuration.
73+
1. Then, review the pricing breakdown for the NIM deployment, terms of use and license agreement associated with the NIM offer. The pricing breakdown helps to inform what the aggregated pricing for the NIM software deployed would be, which is a function of the number of NVIDIA GPUs in the VM instance that was selected in the previous steps. In addition to the applicable NIM software price, Azure Compute charges also applies based on your deployment configuration.
7474

75-
:::image type="content" source="../media/how-to/deploy-nvidia-inference-microservice/payment-description.png" alt-text="A screenshot showing the necessary user payment agreement detailing how the user will be charged for deploying the models." lightbox="../media/how-to/deploy-nvidia-inference-microservice/payment-description.png":::
75+
:::image type="content" source="../media/how-to/deploy-nvidia-inference-microservice/payment-description.png" alt-text="A screenshot showing the necessary user payment agreement detailing how the user is charged for deploying the models." lightbox="../media/how-to/deploy-nvidia-inference-microservice/payment-description.png":::
7676

77-
1. Click the checkbox to acknowledge understanding of pricing and terms of use, and then, click **Deploy**.
77+
1. Select the checkbox to acknowledge understanding of pricing and terms of use, and then, select **Deploy**.
7878

7979
:::image type="content" source="../media/how-to/deploy-nvidia-inference-microservice/deploy-nvidia-inference-microservice.png" alt-text="A screenshot showing the deploy model button in the deployment wizard." lightbox="../media/how-to/deploy-nvidia-inference-microservice/deploy-nim.png":::
8080

8181

8282
## Consume NVIDIA NIM deployments
8383

84-
After your deployment has been successfully created, you can go to **Models + Endpoints** under My assets in your AI Foundry project, select your deployment under "Model deployments" and navigate to the Test tab for sample inference to the endpoint. You can also go to the Chat Playground by clicking **Open in Playground** in Deployment Details tab, to be able to modify parameters for the inference requests.
84+
After your deployment is successfully created, you can go to **Models + Endpoints** under My assets in your Azure AI Foundry project, select your deployment under "Model deployments" and navigate to the Test tab for sample inference to the endpoint. You can also go to the Chat Playground by selecting **Open in Playground** in Deployment Details tab, to be able to modify parameters for the inference requests.
8585

8686
NVIDIA NIMs on Foundry expose an OpenAI compatible API, learn more about the payload supported [here](https://docs.nvidia.com/nim/large-language-models/latest/api-reference.html#). The 'model' parameter for NIMs on Foundry is set to a default value within the container, and is not required to pass through in the payload to your online endpoint. The **Consume** tab of the NIM deployment on Foundry includes code samples for inference with the target URL of your deployment. You can also consume NIM deployments using the Azure AI Model Inference SDK.
8787

@@ -92,8 +92,8 @@ Redeploy to get the latest version of NIM from NVIDIA on Foundry.
9292

9393
## Network Isolation support for NIMs
9494

95-
NVIDIA ensures the security and reliability of NVIDIA NIM container images through best-in-class vulnerability scanning, rigorous patch management, and transparent processes. Learn the details [here](https://docs.nvidia.com/ai-enterprise/planning-resource/security-for-azure-ai-foundry/latest/introduction.html). Microsoft works with NVIDIA to get the latest patches of the NIMs to deliver secure, stable, and reliable production-grade software within Azure AI Foundry.
96-
Users can refer to the last updated time for the NIM in the model overview page, and will have to redeploy to get the latest version of NIM from NVIDIA on Foundry.
95+
NVIDIA ensures the security and reliability of NVIDIA NIM container images through best-in-class vulnerability scanning, rigorous patch management, and transparent processes. Learn the details [here](https://docs.nvidia.com/ai-enterprise/planning-resource/security-for-azure-ai-foundry/latest/introduction.html). Microsoft works with NVIDIA to get the latest patches of the NIMs to deliver secure, stable, and reliable production-grade software within AI Foundry.
96+
Users can refer to the last updated time for the NIM in the model overview page, and you can redeploy to get the latest version of NIM from NVIDIA on Foundry.
9797

9898
While NIMs are in preview on Foundry, workspaces with Public Network Access disabled will have a limitation of being able to create only one successful deployment in the private workspace or project. Note, there can only be a single active deployment in a private workspace, attempts to create more active deployments will end in failure.
9999

0 commit comments

Comments
 (0)