Skip to content

Commit 9b7a353

Browse files
authored
Merge pull request #49401 from ShawnKupfer/WB1695
User Story 399577: Q&M: Retire path and modules: NVidia + ML (2 of 2)
2 parents c8a1600 + 7cccee6 commit 9b7a353

File tree

62 files changed

+80
-771
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

62 files changed

+80
-771
lines changed

.openpublishing.redirection.json

Lines changed: 70 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -44630,6 +44630,76 @@
4463044630
"source_path": "learn-pr/nvidia/create-workspace-resources-getting-started-azure-machine-learning/index.yml",
4463144631
"redirect_url": "https://learn.microsoft.com/training/browse/",
4463244632
"redirect_document_id": false
44633+
},
44634+
{
44635+
"source_path": "learn-pr/nvidia/deploy-model-to-nvidia-triton-inference-server/1-introduction.yml",
44636+
"redirect_url": "https://learn.microsoft.com/training/browse/",
44637+
"redirect_document_id": false
44638+
},
44639+
{
44640+
"source_path": "learn-pr/nvidia/deploy-model-to-nvidia-triton-inference-server/2-create-gpu-accelerated-virtual-machine.yml",
44641+
"redirect_url": "https://learn.microsoft.com/training/browse/",
44642+
"redirect_document_id": false
44643+
},
44644+
{
44645+
"source_path": "learn-pr/nvidia/deploy-model-to-nvidia-triton-inference-server/3-install-prerequisites-nvidia-triton-inference-server.yml",
44646+
"redirect_url": "https://learn.microsoft.com/training/browse/",
44647+
"redirect_document_id": false
44648+
},
44649+
{
44650+
"source_path": "learn-pr/nvidia/deploy-model-to-nvidia-triton-inference-server/4-execute-inference-workload-nvidia-triton-inference-server.yml",
44651+
"redirect_url": "https://learn.microsoft.com/training/browse/",
44652+
"redirect_document_id": false
44653+
},
44654+
{
44655+
"source_path": "learn-pr/nvidia/deploy-model-to-nvidia-triton-inference-server/5-knowledge-check.yml",
44656+
"redirect_url": "https://learn.microsoft.com/training/browse/",
44657+
"redirect_document_id": false
44658+
},
44659+
{
44660+
"source_path": "learn-pr/nvidia/deploy-model-to-nvidia-triton-inference-server/6-summary.yml",
44661+
"redirect_url": "https://learn.microsoft.com/training/browse/",
44662+
"redirect_document_id": false
44663+
},
44664+
{
44665+
"source_path": "learn-pr/nvidia/deploy-model-to-nvidia-triton-inference-server/index.yml",
44666+
"redirect_url": "https://learn.microsoft.com/training/browse/",
44667+
"redirect_document_id": false
44668+
},
44669+
{
44670+
"source_path": "learn-pr/nvidia/use-automl-train-labeled-dataset-develop-production-model/1-introduction.yml",
44671+
"redirect_url": "https://learn.microsoft.com/training/browse/",
44672+
"redirect_document_id": false
44673+
},
44674+
{
44675+
"source_path": "learn-pr/nvidia/use-automl-train-labeled-dataset-develop-production-model/2-prepare-jupyter-notebook-workspace.yml",
44676+
"redirect_url": "https://learn.microsoft.com/training/browse/",
44677+
"redirect_document_id": false
44678+
},
44679+
{
44680+
"source_path": "learn-pr/nvidia/use-automl-train-labeled-dataset-develop-production-model/3-configure-jupyter-notebook-execution-environment.yml",
44681+
"redirect_url": "https://learn.microsoft.com/training/browse/",
44682+
"redirect_document_id": false
44683+
},
44684+
{
44685+
"source_path": "learn-pr/nvidia/use-automl-train-labeled-dataset-develop-production-model/4-execute-jupyter-notebook-produce-object-detection-model-automl.yml",
44686+
"redirect_url": "https://learn.microsoft.com/training/browse/",
44687+
"redirect_document_id": false
44688+
},
44689+
{
44690+
"source_path": "learn-pr/nvidia/use-automl-train-labeled-dataset-develop-production-model/5-knowledge-check.yml",
44691+
"redirect_url": "https://learn.microsoft.com/training/browse/",
44692+
"redirect_document_id": false
44693+
},
44694+
{
44695+
"source_path": "learn-pr/nvidia/use-automl-train-labeled-dataset-develop-production-model/6-summary.yml",
44696+
"redirect_url": "https://learn.microsoft.com/training/browse/",
44697+
"redirect_document_id": false
44698+
},
44699+
{
44700+
"source_path": "learn-pr/nvidia/use-automl-train-labeled-dataset-develop-production-model/index.yml",
44701+
"redirect_url": "https://learn.microsoft.com/training/browse/",
44702+
"redirect_document_id": false
4463344703
}
4463444704
]
4463544705
}

learn-pr/achievements.yml

Lines changed: 10 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -5325,6 +5325,16 @@ achievements:
53255325
title: Write automated tests
53265326
summary: Learn about the types of testing you can do with Rust.
53275327
iconUrl: /training/achievements/rust-automated-tests.svg
5328+
- uid: learn.nvidia.use-automl-train-labeled-dataset-develop-production-model.badge
5329+
type: badge
5330+
title: Use AutoML to train a labeled dataset and develop a production model
5331+
summary: Learn how to use Automated Machine Learning to train a labeled dataset and develop a production object detection model.
5332+
iconUrl: /training/achievements/setup-configure-nvidia-deepstream-development.svg
5333+
- uid: learn.nvidia.deploy-model-to-nvidia-triton-inference-server.badge
5334+
type: badge
5335+
title: Deploy model to NVIDIA Triton Inference Server
5336+
summary: NVIDIA Triton Inference Server is a multi-framework, open-source software that is optimized for inference. It supports popular machine learning frameworks like TensorFlow, Open Neural Network Exchange (ONNX) Runtime, PyTorch, NVIDIA TensorRT, and more. It can be used for your CPU or GPU workloads. In this module, you deploy your production model to NVIDIA Triton server to perform inference on a cloud-hosted virtual machine.
5337+
iconUrl: /training/achievements/introduction-nvidia-deepstream-graph-composer-azure.svg
53285338
- uid: learn.nvidia.develop-custom-object-detection-models-with-nvidia-and-azure-machine-learning.trophy
53295339
type: trophy
53305340
title: Develop Custom Object Detection Models with NVIDIA and Azure Machine Learning

learn-pr/nvidia/deploy-model-to-nvidia-triton-inference-server/1-introduction.yml

Lines changed: 0 additions & 17 deletions
This file was deleted.

learn-pr/nvidia/deploy-model-to-nvidia-triton-inference-server/2-create-gpu-accelerated-virtual-machine.yml

Lines changed: 0 additions & 17 deletions
This file was deleted.

learn-pr/nvidia/deploy-model-to-nvidia-triton-inference-server/3-install-prerequisites-nvidia-triton-inference-server.yml

Lines changed: 0 additions & 17 deletions
This file was deleted.

learn-pr/nvidia/deploy-model-to-nvidia-triton-inference-server/4-execute-inference-workload-nvidia-triton-inference-server.yml

Lines changed: 0 additions & 17 deletions
This file was deleted.

learn-pr/nvidia/deploy-model-to-nvidia-triton-inference-server/5-knowledge-check.yml

Lines changed: 0 additions & 52 deletions
This file was deleted.

learn-pr/nvidia/deploy-model-to-nvidia-triton-inference-server/6-summary.yml

Lines changed: 0 additions & 17 deletions
This file was deleted.

learn-pr/nvidia/deploy-model-to-nvidia-triton-inference-server/includes/1-introduction.md

Lines changed: 0 additions & 21 deletions
This file was deleted.

learn-pr/nvidia/deploy-model-to-nvidia-triton-inference-server/includes/2-create-gpu-accelerated-virtual-machine.md

Lines changed: 0 additions & 59 deletions
This file was deleted.

0 commit comments

Comments
 (0)