-
Notifications
You must be signed in to change notification settings - Fork 749
NXP backend: Replace pass to fuse activations functions with joint quantization with activation #14816
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/14816
Note: Links to docs will display an error until the docs builds have been completed. ❌ 2 New FailuresAs of commit 6f74f12 with merge base 1076686 ( NEW FAILURES - The following jobs have failed:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
|
@pytorchbot label "module: nxp" "release notes: nxp" |
33f153a to
8d9107b
Compare
StrycekSimon
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Please remove the comments about shared quantization if it is not relevant.
…rch_pipeline.py, fix edge pass manager usage
796789f to
ce107df
Compare
+ Move fused activations to separate QDQ cluster
ce107df to
6f74f12
Compare
Summary
This PR replaces optimizations 'fuse_activation_functions.py' by quantization of Conv 2D and Linear ops together with fusable activations - selected activations supported by Neutron (Relu, Relu6, Sigmoid, Tanh). Logic is determined by target specs, now supporting Neutron-C. Tests updated. Relu has now non-shared, standalone quantization.
Test plan
Unit tests provided (test_edge_passes.py, test_quantizer.py).
cc @robert-kalmar @JakeStevens @digantdesai