You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: content/learning-paths/servers-and-cloud-computing/onnx-on-azure/baseline.md
+10-8Lines changed: 10 additions & 8 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -7,12 +7,11 @@ layout: learningpathall
7
7
---
8
8
9
9
10
-
## Baseline testing using ONNX Runtime:
10
+
## Baseline Testing using ONNX Runtime:
11
11
12
-
This test measures the inference latency of the ONNX Runtime by timing how long it takes to process a single input using the `squeezenet-int8.onnx model`. It helps evaluate how efficiently the model runs on the target hardware.
13
-
14
-
Create a **baseline.py** file with the below code for baseline test of ONNX:
12
+
The purpose of this test is to measure the inference latency of ONNX Runtime on your Azure Cobalt 100 VM. By timing how long it takes to process a single input through the SqueezeNet INT8 model, you can validate that ONNX Runtime is functioning correctly and get a baseline performance measurement for your target hardware.
15
13
14
+
Create a file named `baseline.py` with the following code:
16
15
```python
17
16
import onnxruntime as ort
18
17
import numpy as np
@@ -29,12 +28,12 @@ end = time.time()
29
28
print("Inference time:", end - start)
30
29
```
31
30
32
-
Run the baseline test:
31
+
Run the baseline script to measure inference time:
0 commit comments