Skip to content

Commit 22edbac

Browse files
authored
Add Nvidia collections image to documentation
1 parent 4fc1f45 commit 22edbac

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

articles/ai-foundry/how-to/deploy-nvidia-inference-microservice.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -59,6 +59,9 @@ Get improved TCO (total cost of ownership) and performance with NVIDIA NIMs offe
5959
1. Sign in to [Azure AI Foundry](https://ai.azure.com) and go to the **Home** page.
6060
2. Select **Model catalog** from the left sidebar.
6161
3. In the filters section, select **Collections** and select **NVIDIA**.
62+
63+
:::image type="content" source="../media/how-to/deploy-nvidia-inference-microservice/nvidia-collections.png" alt-text="A screenshot showing the Nvidia inference microservices available in the model catalog." lightbox="../media/how-to/deploy-nvidia-inference-microservice/nvidia-collections.png":::
64+
6265
4. Select the NVIDIA NIM of your choice. In this article, we are using **Llama-3.3-70B-Instruct-NIM-microservice** as an example.
6366
5. Select **Deploy**.
6467
6. Select one of the NVIDIA GPU based VM SKUs supported for the NIM, based on your intended workload. You need to have quota in your Azure subscription.
@@ -73,9 +76,6 @@ Get improved TCO (total cost of ownership) and performance with NVIDIA NIMs offe
7376

7477
10. Select the checkbox to acknowledge understanding of pricing and terms of use, and then, select **Deploy**.
7578

76-
:::image type="content" source="../media/how-to/deploy-nvidia-inference-microservice/deploy-nvidia-inference-microservice.png" alt-text="A screenshot showing the deploy model button in the deployment wizard." lightbox="../media/how-to/deploy-nvidia-inference-microservice/deploy-nvidia-inference-microservice.png":::
77-
78-
7979
## Consume NVIDIA NIM deployments
8080

8181
After your deployment is successfully created, you can go to **Models + Endpoints** under My assets in your Azure AI Foundry project, select your deployment under "Model deployments" and navigate to the Test tab for sample inference to the endpoint. You can also go to the Chat Playground by selecting **Open in Playground** in Deployment Details tab, to be able to modify parameters for the inference requests.

0 commit comments

Comments
 (0)