Skip to content

Commit 73d74ca

Browse files
committed
toc, nits
1 parent 00bf857 commit 73d74ca

File tree

2 files changed

+5
-5
lines changed

2 files changed

+5
-5
lines changed

articles/container-apps/TOC.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -281,7 +281,7 @@
281281
- name: Generate images with serverless GPUs
282282
href: gpu-image-generation.md
283283
- name: Deploy an NVIDIA Llama3 NIM
284-
href: tutorial-gpu-with-serverless-gpu.md
284+
href: serverless-gpu-nim.md
285285
- name: Microservices
286286
items:
287287
- name: Developing with Dapr

articles/container-apps/tutorial-gpu-with-serverless-gpu.md renamed to articles/container-apps/serverless-gpu-nim.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ ms.devlang: azurecli
1212

1313
# Tutorial: Deploy an NVIDIA Llama3 NIM to Azure Container Apps
1414

15-
NVIDIA Inference Microservices (NIMs) are optimized, containerized AI inference microservices designed to simplify and accelerate the deployment of AI models. When you use Azure Container Apps with serverless GPUs, you can run these NIMs efficiently without having to manage the underlying infrastructure.​
15+
NVIDIA Inference Microservices (NIMs) are optimized, containerized AI inference microservices which simplify and accelerate the development of AI applications and agentic AI workflows with pre-packaged, scalable, and performance-tuned models that can be deployed as secure inference endpoints on Azure Container Apps. When you use Azure Container Apps with serverless GPUs, you can run these NIMs efficiently without having to manage the underlying infrastructure.​
1616

1717
In this tutorial, you learn to deploy a Llama3 NVIDIA NIM to Azure Container Apps using serverless GPUs.
1818

@@ -38,7 +38,7 @@ This tutorial uses a premium instance of Azure Container Registry to improve col
3838
Next, generate a unique container registry name.
3939

4040
```bash
41-
SUFFIX=$(head /dev/urandom | tr -dc 'A-Za-z0-9' | head -c 6)
41+
SUFFIX=$(head /dev/urandom | tr -dc 'a-z0-9' | head -c 6)
4242
ACR_NAME="mygpututorialacr${SUFFIX}"
4343
```
4444

@@ -72,7 +72,7 @@ This tutorial uses a premium instance of Azure Container Registry to improve col
7272
Next, pull the image from NVIDIA GPU Cloud and push to Azure Container Registry.
7373
7474
> [!NOTE]
75-
> NVIDIA NICs each has their own hardware requirements. Make sure the GPU type you select supports the [NIM](link) of your choice. The Llama3 NIM used in this tutorial can run on NVIDIA A100 GPUs.
75+
> NVIDIA NICs each has their own hardware requirements. Make sure the GPU type you select supports the [NIM](https://build.nvidia.com/models?filters=nimType%3Anim_type_run_anywhere&q=llama) of your choice. The Llama3 NIM used in this tutorial can run on NVIDIA A100 GPUs.
7676
7777
1. Authenticate to the NVIDIA container registry.
7878
@@ -110,7 +110,7 @@ Next, pull the image from NVIDIA GPU Cloud and push to Azure Container Registry.
110110

111111
## Enable artifact streaming (recommended but optional)
112112

113-
Many of the NIM images are large, and your container app can take a long time to start if you don't enable artifact streaming. Use the following steps to enable artifact streaming.
113+
When your container app runs, it pulls the container from your container registry. When you have larger images like in the case of AI workloads, this image pull may take some time. By enabling artifact streaming, you reduce the time needed , and your container app can take a long time to start if you don't enable artifact streaming. Use the following steps to enable artifact streaming.
114114
115115
> [!NOTE]
116116
> The following commands can take a few minutes to complete.

0 commit comments

Comments
 (0)