Questions about using XNNPACK Execution Provider #18112
Unanswered
postech-sms
asked this question in
Other Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi, all
I have questions about using XNNPACK Execution Provider. (linux-arm64 ubuntu:20.04)
I built onnxruntime including xnnpack from the v1.16.0 tag in the onnxruntime repository. (use --use_xnnpack)
Using my built files,
I checked that XNNExecutionProvider works well For the Full Precision Model.
(I mean that convolution layers are mapped into XNNExecutionProvider)
However, when I converted Full Precision Model to INT8 Quantization Model,
Convolution layers weren't mapped into XNNExecutionProvider
(Just mapped into CPUExecutionProvider)
I confirmed in the official documentation that XNNExecutionProvider also supports QLinearConv. (ai.onnx:QLinearConv)
(https://github.com/microsoft/onnxruntime/blob/gh-pages/docs/execution-providers/Xnnpack-ExecutionProvider.md#supported-ops)
But I don't understand why XNNPackExecutionProvider doesn't work properly in my INT8 model.
This may be an issue that occurred because I am a newbie, so please understand.
Thanks.
Beta Was this translation helpful? Give feedback.
All reactions