Skip to content

Commit 325a951

Browse files
committed
edits
1 parent f8db016 commit 325a951

File tree

1 file changed

+14
-14
lines changed

1 file changed

+14
-14
lines changed

articles/machine-learning/concept-onnx.md

Lines changed: 14 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
---
2-
title: ONNX Runtime and models
2+
title: ONNX Runtime and Models
33
titleSuffix: Azure Machine Learning
44
description: Learn how using the Open Neural Network Exchange (ONNX) can help optimize inference of your machine learning models.
55
services: machine-learning
@@ -9,15 +9,15 @@ ms.topic: concept-article
99
ms.author: scottpolly
1010
author: s-polly
1111
ms.reviewer: kritifaujdar
12-
ms.date: 09/30/2024
12+
ms.date: 10/13/2025
1313
#customer intent: As a data scientist, I want to learn about ONNX so I can use it to optimize the inference of my machine learning models.
1414
---
1515

1616
# ONNX and Azure Machine Learning
1717

18-
This article describes how the [Open Neural Network Exchange (ONNX)](https://onnx.ai) can help optimize the *inference* of your machine learning models. Inference or model scoring is the process of using a deployed model to generate predictions on production data.
18+
This article describes how the [Open Neural Network Exchange (ONNX)](https://onnx.ai) can help optimize the inference of your machine learning models. *Inference*, or model scoring, is the process of using a deployed model to generate predictions on production data.
1919

20-
Optimizing machine learning models for inference requires you to tune the model and the inference library to make the most of hardware capabilities. This task becomes complex if you want to get optimal performance on different platforms such as cloud, edge, CPU, or GPU, because each platform has different capabilities and characteristics. The complexity increases if you need to run models from various frameworks on different platforms. It can be time-consuming to optimize all the different combinations of frameworks and hardware.
20+
Optimizing machine learning models for inference requires you to tune the model and the inference library to make the most of hardware capabilities. This task becomes complex if you want to get optimal performance on different platforms, such as cloud, edge, CPU, or GPU, because each platform has different capabilities and characteristics. The complexity increases if you need to run models from various frameworks on different platforms. It can be time-consuming to optimize all the different combinations of frameworks and hardware.
2121

2222
A useful solution is to train your model one time in your preferred framework, and then export or convert it to ONNX so it can run anywhere on the cloud or edge. Microsoft and a community of partners created ONNX as an open standard for representing machine learning models. You can export or convert models from [many frameworks](https://onnx.ai/supported-tools) to the standard ONNX format. Supported frameworks include TensorFlow, PyTorch, scikit-learn, Keras, Chainer, MXNet, and MATLAB. You can run models in the ONNX format on various platforms and devices.
2323

@@ -27,34 +27,34 @@ This ONNX flow diagram shows available frameworks and deployment options.
2727

2828
## ONNX Runtime
2929

30-
[ONNX Runtime](https://onnxruntime.ai) is a high-performance inference engine for deploying ONNX models to production. ONNX Runtime is optimized for both cloud and edge, and works on Linux, Windows, and macOS. ONNX is written in C++, but also has C, Python, C#, Java, and JavaScript (Node.js) APIs to use in those environments.
30+
[ONNX Runtime](https://onnxruntime.ai) is a high-performance inference engine for deploying ONNX models to production. ONNX Runtime is optimized for both cloud and edge, and works on Linux, Windows, and macOS. ONNX is written in C++, but also has C, Python, C#, Java, and JavaScript (Node.js) APIs that you can use in those environments.
3131

32-
ONNX Runtime supports both deep neural networks (DNN) and traditional machine learning models, and it integrates with accelerators on different hardware such as TensorRT on NVIDIA GPUs, OpenVINO on Intel processors, and DirectML on Windows. By using ONNX Runtime, you can benefit from extensive production-grade optimizations, testing, and ongoing improvements.
32+
ONNX Runtime supports both deep neural networks (DNN) and traditional machine learning models. It integrates with accelerators on different hardware, such as TensorRT on NVIDIA GPUs, OpenVINO on Intel processors, and DirectML on Windows. By using ONNX Runtime, you can benefit from extensive production-grade optimizations, testing, and ongoing improvements.
3333

34-
High-scale Microsoft services such as Bing, Office, and Azure AI use ONNX Runtime. Although performance gains depend on many factors, these Microsoft services report an average 2x performance gain on CPU by using ONNX. ONNX Runtime runs in Azure Machine Learning and other Microsoft products that support machine learning workloads, including:
34+
High-scale Microsoft services such as Bing, Office, and Azure AI use ONNX Runtime. Although performance gains depend on many factors, these Microsoft services average a 2x performance gain on CPU because they use ONNX. ONNX Runtime runs in Azure Machine Learning and other Microsoft products that support machine learning workloads, including:
3535

36-
- **Windows**. ONNX runtime is built into Windows as part of [Windows Machine Learning](/windows/ai/windows-ml/) and runs on hundreds of millions of devices.
36+
- **Windows**. ONNX Runtime is built into Windows as part of [Windows Machine Learning](/windows/ai/windows-ml/) and runs on hundreds of millions of devices.
3737
- **Azure SQL**. [Azure SQL Edge](/azure/azure-sql-edge/onnx-overview) and [Azure SQL Managed Instance](/azure/azure-sql/managed-instance/machine-learning-services-overview) use ONNX to run native scoring on data.
38-
- **ML.NET**. For an example, see [Tutorial: Detect objects using ONNX in ML.NET](/dotnet/machine-learning/tutorials/object-detection-onnx).
38+
- **ML.NET**. For an example, see [Tutorial: Detect objects by using ONNX in ML.NET](/dotnet/machine-learning/tutorials/object-detection-onnx).
3939

40-
## Ways to obtain ONNX models
40+
## Ways to get ONNX models
4141

42-
You can obtain ONNX models in several ways:
42+
You can get ONNX models in several ways:
4343

4444
- [Train a new ONNX model in Azure Machine Learning](https://github.com/onnx/onnx/tree/main/examples) or use [automated machine learning capabilities](concept-automated-ml.md#automl--onnx).
4545
- Convert an existing model from another format to ONNX. For more information, see [ONNX Tutorials](https://github.com/onnx/tutorials).
4646
- Get a pretrained ONNX model from the [ONNX Model Zoo](https://github.com/onnx/models).
47-
- Generate a customized ONNX model from [Azure AI Custom Vision service](/azure/ai-services/custom-vision-service/).
47+
- Generate a customized ONNX model from the [Azure AI Custom Vision service](/azure/ai-services/custom-vision-service/).
4848

4949
You can represent many models as ONNX, including image classification, object detection, and text processing models. If you can't convert your model successfully, file a GitHub issue in the repository of the converter you used.
5050

5151
## ONNX model deployment in Azure
5252

53-
You can deploy, manage, and monitor your ONNX models in Azure Machine Learning. Using a standard [MLOps deployment workflow](concept-model-management-and-deployment.md#deploy-models-as-endpoints) with ONNX Runtime, you can create a REST endpoint hosted in the cloud.
53+
You can deploy, manage, and monitor your ONNX models in Azure Machine Learning. By using a standard [MLOps deployment workflow](concept-model-management-and-deployment.md#deploy-models-as-endpoints) with ONNX Runtime, you can create a REST endpoint hosted in the cloud.
5454

5555
## Python packages for ONNX Runtime
5656

57-
Python packages for [CPU](https://pypi.org/project/onnxruntime) and [GPU](https://pypi.org/project/onnxruntime-gpu) ONNX Runtime are available on [PyPi.org](https://pypi.org). Be sure to review system requirements before installation.
57+
Python packages for the [CPU](https://pypi.org/project/onnxruntime) and [GPU](https://pypi.org/project/onnxruntime-gpu) ONNX Runtime are available on [PyPi.org](https://pypi.org). Be sure to review system requirements before installation.
5858

5959
To install ONNX Runtime for Python, use one of the following commands:
6060

0 commit comments

Comments
 (0)