Accuracy issues moving from AIMET-ONNX to SNPE DLC #4010
-
|
Hi AIMET team, I’m experiencing a big accuracy gap between the QDQ ONNX model exported (through the to_onnx_qdq()) and tested with onnxruntime, and the DLC I generate with SNPE using the I am trying to quantize a pose estimation model (Lite-HRNet) and I am visually evaluating the results on a set of images. Following the exact same pre and post processing steps I am getting almost identical results between the FP32 ONNX, the QDQ ONNX (produced by AIMET), and the FP16 converted DLC. But when I try the INT8 quantized DLC (converted using the .onnx and .encodings from AIMET) the results are way off. Here is the pipeline I’m following:
Things I have already tried
Questions
I have searched through existing discussions and issues, but most similar posts involve PyTorch models instead of ONNX or use QNN rather than SNPE for deployment; I have tried the suggestions from those threads without success, so I’m opening this new discussion. |
Beta Was this translation helpful? Give feedback.
Replies: 3 comments 6 replies
-
|
Hi @ptoupas - Thanks for reaching out to us. And apologies for a delayed response. |
Beta Was this translation helpful? Give feedback.
-
|
So here is the support forums link for SNPE (Qualcomm Neural Processing SDK), QNN (Neural Network SDK), QAIRT etc. Note: Please don't use the above for AIMET questions. Just create discussion threads here for AIMET. |
Beta Was this translation helpful? Give feedback.
-
|
Hi |
Beta Was this translation helpful? Give feedback.
So here is the support forums link for SNPE (Qualcomm Neural Processing SDK), QNN (Neural Network SDK), QAIRT etc.
https://mysupport.qualcomm.com/supportforums/s/topic/0TO4V0000001XL9WAM/ai-sdks-tools
Note: Please don't use the above for AIMET questions. Just create discussion threads here for AIMET.