Replies: 1 comment 1 reply
-
Experiencing a similar issue when running Pytorch vs. ONNX on Mac M2... curious what's going on here. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I am testing inference performance between
roberta-base
original model vs onnx on CPU (Apple MacBook M1/16G RAM). From the test, it shows that while the sequence length is increasing, ONNX model inference is getting slower comparing to the original PyTorch model. The example code I can find from googling were mostly working with short sequences which demos ONNX is faster than original PyTorch model but no tests I have seen comparing different sequence lengths all the way to the maximum.Is this expected or something I need to do to get ONNX running faster? Or if I did anything wrong, please also let me know. Thanks for any comments. I have attached my code and package versions in the below.
Beta Was this translation helpful? Give feedback.
All reactions