Skip to content

Commit 19b4fbc

Browse files
authored
Merge pull request #203 from changeworld/patch-1
Fix typo
2 parents 7d88a5f + 18b30b7 commit 19b4fbc

14 files changed

+21
-21
lines changed

articles/ai-services/computer-vision/spatial-analysis-container.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -159,9 +159,9 @@ Follow these steps to remotely connect from a Windows client.
159159
160160
Follow these instructions if your host computer isn't an Azure Stack Edge device.
161161
162-
#### Install NVIDIA CUDA Toolkit and Nvidia graphics drivers on the host computer
162+
#### Install NVIDIA CUDA Toolkit and NVIDIA graphics drivers on the host computer
163163
164-
Use the following bash script to install the required Nvidia graphics drivers, and CUDA Toolkit.
164+
Use the following bash script to install the required NVIDIA graphics drivers, and CUDA Toolkit.
165165
166166
```bash
167167
wget https://developer.download.nvidia.com/compute/cuda/repos/ubuntu1804/x86_64/cuda-ubuntu1804.pin

articles/ai-services/content-safety/how-to/containers/install-run-container.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -86,7 +86,7 @@ Even with identical GPUs, performance can fluctuate based on the GPU load and th
8686

8787
## Install the NVIDIA container toolkit
8888

89-
The `host` is the computer that runs the docker container. The host must support Nvidia container toolkit. Follow the below guidance to install the toolkit in your environment.
89+
The `host` is the computer that runs the docker container. The host must support NVIDIA container toolkit. Follow the below guidance to install the toolkit in your environment.
9090

9191
[Install the NVIDIA Container Toolkit](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html)
9292

articles/ai-services/language-service/summarization/how-to/use-containers.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@ The following table describes the minimum and recommended specifications for the
3535
| Container Type | Recommended number of CPU cores | Recommended memory | Notes |
3636
|----------------------------|----------------------------------|--------------------|-------|
3737
| Summarization CPU container| 16 | 48 GB | |
38-
| Summarization GPU container| 2 | 24 GB | Requires an Nvidia GPU that supports Cuda 11.8 with 16GB VRAM.|
38+
| Summarization GPU container| 2 | 24 GB | Requires an NVIDIA GPU that supports Cuda 11.8 with 16GB VRAM.|
3939

4040
CPU core and memory correspond to the `--cpus` and `--memory` settings, which are used as part of the `docker run` command.
4141

articles/ai-studio/how-to/deploy-models-mistral-nemo.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ Mistral Nemo is a 12B model, making it a powerful drop-in replacement for any sy
3737

3838
Additionally, Mistral Nemo is:
3939

40-
* **Jointly developed with Nvidia**. This collaboration has resulted in a powerful 12B model that pushes the boundaries of language understanding and generation.
40+
* **Jointly developed with NVIDIA**. This collaboration has resulted in a powerful 12B model that pushes the boundaries of language understanding and generation.
4141
* **Multilingual proficient**. Mistral Nemo is equipped with a tokenizer called Tekken, which is designed for multilingual applications. It supports over 100 languages, such as English, French, German, and Spanish. Tekken is more efficient than the Llama 3 tokenizer in compressing text for approximately 85% of all languages, with significant improvements in Malayalam, Hindi, Arabic, and prevalent European languages.
4242
* **Agent-centric**. Mistral Nemo possesses top-tier agentic capabilities, including native function calling and JSON outputting.
4343
* **Advanced in reasoning**. Mistral Nemo demonstrates state-of-the-art mathematical and reasoning capabilities within its size category.
@@ -477,7 +477,7 @@ Mistral Nemo is a 12B model, making it a powerful drop-in replacement for any sy
477477

478478
Additionally, Mistral Nemo is:
479479

480-
* **Jointly developed with Nvidia**. This collaboration has resulted in a powerful 12B model that pushes the boundaries of language understanding and generation.
480+
* **Jointly developed with NVIDIA**. This collaboration has resulted in a powerful 12B model that pushes the boundaries of language understanding and generation.
481481
* **Multilingual proficient**. Mistral Nemo is equipped with a tokenizer called Tekken, which is designed for multilingual applications. It supports over 100 languages, such as English, French, German, and Spanish. Tekken is more efficient than the Llama 3 tokenizer in compressing text for approximately 85% of all languages, with significant improvements in Malayalam, Hindi, Arabic, and prevalent European languages.
482482
* **Agent-centric**. Mistral Nemo possesses top-tier agentic capabilities, including native function calling and JSON outputting.
483483
* **Advanced in reasoning**. Mistral Nemo demonstrates state-of-the-art mathematical and reasoning capabilities within its size category.
@@ -936,7 +936,7 @@ Mistral Nemo is a 12B model, making it a powerful drop-in replacement for any sy
936936
937937
Additionally, Mistral Nemo is:
938938
939-
* **Jointly developed with Nvidia**. This collaboration has resulted in a powerful 12B model that pushes the boundaries of language understanding and generation.
939+
* **Jointly developed with NVIDIA**. This collaboration has resulted in a powerful 12B model that pushes the boundaries of language understanding and generation.
940940
* **Multilingual proficient**. Mistral Nemo is equipped with a tokenizer called Tekken, which is designed for multilingual applications. It supports over 100 languages, such as English, French, German, and Spanish. Tekken is more efficient than the Llama 3 tokenizer in compressing text for approximately 85% of all languages, with significant improvements in Malayalam, Hindi, Arabic, and prevalent European languages.
941941
* **Agent-centric**. Mistral Nemo possesses top-tier agentic capabilities, including native function calling and JSON outputting.
942942
* **Advanced in reasoning**. Mistral Nemo demonstrates state-of-the-art mathematical and reasoning capabilities within its size category.
@@ -1417,7 +1417,7 @@ Mistral Nemo is a 12B model, making it a powerful drop-in replacement for any sy
14171417
14181418
Additionally, Mistral Nemo is:
14191419
1420-
* **Jointly developed with Nvidia**. This collaboration has resulted in a powerful 12B model that pushes the boundaries of language understanding and generation.
1420+
* **Jointly developed with NVIDIA**. This collaboration has resulted in a powerful 12B model that pushes the boundaries of language understanding and generation.
14211421
* **Multilingual proficient**. Mistral Nemo is equipped with a tokenizer called Tekken, which is designed for multilingual applications. It supports over 100 languages, such as English, French, German, and Spanish. Tekken is more efficient than the Llama 3 tokenizer in compressing text for approximately 85% of all languages, with significant improvements in Malayalam, Hindi, Arabic, and prevalent European languages.
14221422
* **Agent-centric**. Mistral Nemo possesses top-tier agentic capabilities, including native function calling and JSON outputting.
14231423
* **Advanced in reasoning**. Mistral Nemo demonstrates state-of-the-art mathematical and reasoning capabilities within its size category.

articles/machine-learning/azure-machine-learning-ci-image-release-notes.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ gnomeshell: removed
3636

3737
ml: '2.32.4'
3838

39-
Nvidia Driver: `535.216.03`
39+
NVIDIA Driver: `535.216.03`
4040

4141
`CUDA`: `12.2`
4242

@@ -50,7 +50,7 @@ SDK Version: `1.57.0`
5050

5151
Python: `3.10.11`
5252

53-
Nvidia Driver: `535.183.06`
53+
NVIDIA Driver: `535.183.06`
5454

5555
`CUDA`: `12.2`
5656

@@ -223,7 +223,7 @@ Python: `3.9`
223223

224224
CUDnn==`9.1.1`
225225

226-
Nvidia Driver: `535.171.04`
226+
NVIDIA Driver: `535.171.04`
227227

228228
PyTorch: `1.13.1`
229229

articles/machine-learning/concept-model-catalog.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ ms.date: 12/11/2024
1717

1818
# Model Catalog and Collections
1919

20-
The model catalog in Azure Machine Learning studio is the hub to discover and use a wide range of models that enable you to build Generative AI applications. The model catalog features hundreds of models from model providers such as Azure OpenAI service, Mistral, Meta, Cohere, Nvidia, Hugging Face, including models trained by Microsoft. Models from providers other than Microsoft are Non-Microsoft Products, as defined in [Microsoft's Product Terms](https://www.microsoft.com/licensing/terms/welcome/welcomepage), and subject to the terms provided with the model.
20+
The model catalog in Azure Machine Learning studio is the hub to discover and use a wide range of models that enable you to build Generative AI applications. The model catalog features hundreds of models from model providers such as Azure OpenAI service, Mistral, Meta, Cohere, NVIDIA, Hugging Face, including models trained by Microsoft. Models from providers other than Microsoft are Non-Microsoft Products, as defined in [Microsoft's Product Terms](https://www.microsoft.com/licensing/terms/welcome/welcomepage), and subject to the terms provided with the model.
2121

2222
## Model Collections
2323

articles/machine-learning/concept-onnx.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@ This ONNX flow diagram shows available frameworks and deployment options.
2929

3030
[ONNX Runtime](https://onnxruntime.ai) is a high-performance inference engine for deploying ONNX models to production. ONNX Runtime is optimized for both cloud and edge, and works on Linux, Windows, and macOS. ONNX is written in C++, but also has C, Python, C#, Java, and JavaScript (Node.js) APIs to use in those environments.
3131

32-
ONNX Runtime supports both deep neural networks (DNN) and traditional machine learning models, and it integrates with accelerators on different hardware such as TensorRT on Nvidia GPUs, OpenVINO on Intel processors, and DirectML on Windows. By using ONNX Runtime, you can benefit from extensive production-grade optimizations, testing, and ongoing improvements.
32+
ONNX Runtime supports both deep neural networks (DNN) and traditional machine learning models, and it integrates with accelerators on different hardware such as TensorRT on NVIDIA GPUs, OpenVINO on Intel processors, and DirectML on Windows. By using ONNX Runtime, you can benefit from extensive production-grade optimizations, testing, and ongoing improvements.
3333

3434
High-scale Microsoft services such as Bing, Office, and Azure AI use ONNX Runtime. Although performance gains depend on many factors, these Microsoft services report an average 2x performance gain on CPU by using ONNX. ONNX Runtime runs in Azure Machine Learning and other Microsoft products that support machine learning workloads, including:
3535

articles/machine-learning/data-science-virtual-machine/overview.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ Additionally, we offer Azure DSVM for PyTorch - an Ubuntu 20.04 image from Azure
3030
- MSCCL​
3131
- ORTMoE​
3232
- Fairscale​
33-
- Nvidia Apex​
33+
- NVIDIA Apex​
3434
- An up-to-date stack with the latest compatible versions of Ubuntu, Python, PyTorch, and CUDA
3535

3636
## Comparison with Azure Machine Learning

articles/machine-learning/data-science-virtual-machine/release-notes.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -518,7 +518,7 @@ Version: `21.08.11`
518518
Primary changes:
519519

520520
- Windows Security update
521-
- Update of Nvidia CuDNN to 8.1.0
521+
- Update of NVIDIA CuDNN to 8.1.0
522522
- Update of Jupyter Lab -to 3.0.16
523523
- Added MLFLow for experiment tracking
524524
- Improvement of stability and minor bug fixes

articles/machine-learning/how-to-deploy-models-mistral.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -90,7 +90,7 @@ Mistral Nemo is a 12B model, making it a powerful drop-in replacement for any sy
9090

9191
Additionally, Mistral Nemo is:
9292

93-
- **Jointly developed with Nvidia.** This collaboration has resulted in a powerful 12B model that pushes the boundaries of language understanding and generation.
93+
- **Jointly developed with NVIDIA.** This collaboration has resulted in a powerful 12B model that pushes the boundaries of language understanding and generation.
9494
- **Multilingual proficient.** Mistral Nemo is equipped with a tokenizer called Tekken, which is designed for multilingual applications. It supports over 100 languages, such as English, French, German, and Spanish. Tekken is more efficient than the Llama 3 tokenizer in compressing text for approximately 85% of all languages, with significant improvements in Malayalam, Hindi, Arabic, and prevalent European languages.
9595
- **Agent-centric.** Mistral Nemo possesses top-tier agentic capabilities, including native function calling and JSON outputting.
9696
- **Advanced in reasoning.** Mistral Nemo demonstrates state-of-the-art mathematical and reasoning capabilities within its size category.

0 commit comments

Comments
 (0)