-
Notifications
You must be signed in to change notification settings - Fork 722
Adding mixed quantization support #14134
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/14134
Note: Links to docs will display an error until the docs builds have been completed. ✅ No FailuresAs of commit 5188058 with merge base 181ed4d ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
|
This pull request was exported from Phabricator. Differential Revision: D81519735 |
This PR needs a
|
1e64ab6 to
789bf47
Compare
Summary: # Context This Diff adds support for mixed quantization operators in Executorch. Now weights and biases can be quantized, while inputs and activations are kept in floating point. # In this diff 1. Op nodes are returned from each pattern matching 2. Dequantize nodes are bypassed if not needed in the final graph. Differential Revision: D81519735
|
This pull request was exported from Phabricator. Differential Revision: D81519735 |
Summary: Pull Request resolved: pytorch#14134 # Context This Diff adds support for mixed quantization operators in Executorch. Now weights and biases can be quantized, while inputs and activations are kept in floating point. # In this diff 1. Op nodes are returned from each pattern matching 2. Dequantize nodes are bypassed if not needed in the final graph. Differential Revision: D81519735
2ffaef2 to
49f3938
Compare
Summary: # Context This Diff adds support for mixed quantization operators in Executorch. Now weights and biases can be quantized, while inputs and activations are kept in floating point. # In this diff 1. Op nodes are returned from each pattern matching 2. Dequantize nodes are bypassed if not needed in the final graph. Reviewed By: skrtskrtfb Differential Revision: D81519735
49f3938 to
569f2e6
Compare
Summary: # Context This Diff adds support for mixed quantization operators in Executorch. Now weights and biases can be quantized, while inputs and activations are kept in floating point. # In this diff 1. Op nodes are returned from each pattern matching 2. Dequantize nodes are bypassed if not needed in the final graph. Reviewed By: skrtskrtfb Differential Revision: D81519735
569f2e6 to
406d706
Compare
Summary: # Context This Diff adds support for mixed quantization operators in Executorch. Now weights and biases can be quantized, while inputs and activations are kept in floating point. # In this diff 1. Op nodes are returned from each pattern matching 2. Dequantize nodes are bypassed if not needed in the final graph. Reviewed By: skrtskrtfb Differential Revision: D81519735
Summary: # Context This Diff adds support for mixed quantization operators in Executorch. Now weights and biases can be quantized, while inputs and activations are kept in floating point. # In this diff 1. Op nodes are returned from each pattern matching 2. Dequantize nodes are bypassed if not needed in the final graph. Reviewed By: skrtskrtfb Differential Revision: D81519735
406d706 to
2c0b42b
Compare
Summary: # Context This Diff adds support for mixed quantization operators in Executorch. Now weights and biases can be quantized, while inputs and activations are kept in floating point. # In this diff 1. Op nodes are returned from each pattern matching 2. Dequantize nodes are bypassed if not needed in the final graph. Reviewed By: skrtskrtfb Differential Revision: D81519735
2c0b42b to
61701f4
Compare
61701f4 to
10689f5
Compare
Summary: # Context This Diff adds support for mixed quantization operators in Executorch. Now weights and biases can be quantized, while inputs and activations are kept in floating point. # In this diff 1. Op nodes are returned from each pattern matching 2. Dequantize nodes are bypassed if not needed in the final graph. Reviewed By: skrtskrtfb Differential Revision: D81519735
10689f5 to
2840bbd
Compare
2840bbd to
12aae49
Compare
Summary: # Context This Diff adds support for mixed quantization operators in Executorch. Now weights and biases can be quantized, while inputs and activations are kept in floating point. # In this diff 1. Op nodes are returned from each pattern matching 2. Dequantize nodes are bypassed if not needed in the final graph. Reviewed By: skrtskrtfb Differential Revision: D81519735
12aae49 to
b556b8d
Compare
Summary: # Context This Diff adds support for mixed quantization operators in Executorch. Now weights and biases can be quantized, while inputs and activations are kept in floating point. # In this diff 1. Op nodes are returned from each pattern matching 2. Dequantize nodes are bypassed if not needed in the final graph. Reviewed By: skrtskrtfb Differential Revision: D81519735
Summary: # Context This Diff adds support for mixed quantization operators in Executorch. Now weights and biases can be quantized, while inputs and activations are kept in floating point. # In this diff 1. Op nodes are returned from each pattern matching 2. Dequantize nodes are bypassed if not needed in the final graph. Reviewed By: skrtskrtfb Differential Revision: D81519735
b556b8d to
ed2bdec
Compare
ed2bdec to
5b7459b
Compare
Summary: # Context This Diff adds support for mixed quantization operators in Executorch. Now weights and biases can be quantized, while inputs and activations are kept in floating point. # In this diff 1. Op nodes are returned from each pattern matching 2. Dequantize nodes are bypassed if not needed in the final graph. Reviewed By: skrtskrtfb Differential Revision: D81519735
Summary: # Context This Diff adds support for mixed quantization operators in Executorch. Now weights and biases can be quantized, while inputs and activations are kept in floating point. # In this diff 1. Op nodes are returned from each pattern matching 2. Dequantize nodes are bypassed if not needed in the final graph. Reviewed By: skrtskrtfb Differential Revision: D81519735
5b7459b to
1f03b87
Compare
1f03b87 to
a2cceb1
Compare
Summary: # Context This Diff adds support for mixed quantization operators in Executorch. Now weights and biases can be quantized, while inputs and activations are kept in floating point. # In this diff 1. Op nodes are returned from each pattern matching 2. Dequantize nodes are bypassed if not needed in the final graph. Reviewed By: skrtskrtfb Differential Revision: D81519735
a2cceb1 to
792d571
Compare
Summary: # Context This Diff adds support for mixed quantization operators in Executorch. Now weights and biases can be quantized, while inputs and activations are kept in floating point. # In this diff 1. Op nodes are returned from each pattern matching 2. Dequantize nodes are bypassed if not needed in the final graph. Reviewed By: skrtskrtfb Differential Revision: D81519735
Summary: # Context This Diff adds support for mixed quantization operators in Executorch. Now weights and biases can be quantized, while inputs and activations are kept in floating point. # In this diff 1. Op nodes are returned from each pattern matching 2. Dequantize nodes are bypassed if not needed in the final graph. Reviewed By: skrtskrtfb, mcremon-meta Differential Revision: D81519735
792d571 to
63ae795
Compare
Summary:
Context
This Diff adds support for mixed quantization operators in Executorch. Now weights and biases can be quantized, while inputs and activations are kept in floating point.
In this diff
Differential Revision: D81519735