Skip to content

Commit 9126734

Browse files
authored
Fix broken link in deploy-nim.md
1 parent b178c6b commit 9126734

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

articles/ai-foundry/how-to/deploy-nim.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ ms.custom: devx-track-azurecli
1616
# How to deploy NVIDIA Inference Microservices
1717

1818
In this article, you learn how to deploy NVIDIA Inference Microservices (NIMs) on Managed Compute in the model catalog on Foundry​. NVIDIA inference microservices are containers built by NVIDIA for optimized pre-trained and customized AI models serving on NVIDIA GPUs​.
19-
Get improved TCO and performance with NVIDIA NIMs offered for one-click deployment on Foundry, with enterprise production-grade software under NVIDIA AI Enterprise license.
19+
Get improved TCO (total cost of ownership) and performance with NVIDIA NIMs offered for one-click deployment on Foundry, with enterprise production-grade software under NVIDIA AI Enterprise license.
2020

2121
[!INCLUDE [models-preview](../includes/models-preview.md)]
2222

0 commit comments

Comments
 (0)