This repository was archived by the owner on Jun 3, 2025. It is now read-only.
Commit 1db70be
Quantization refactor (#663)
* Removed output quantization from conv layers
* Added _Add_ReLU module that enables QATWrapper for quantizaiton.
* Removed quantization of output for linear and conv layers by default. Removed fusing of BN and ReLU by default.
* Minor fixes. Style and quality fixes.
* Added support to freezing bn stats.
* Added mode argument to wrapping of train function in BNWrapper
* Set BN fusing back as default.
* Set BN fusing back as default.
* Fixed custom freeze_bn_stats.
* Temporary files for evaluating changes to graphs.
* Added support to tensorrt flag. Moved the computation of quantization range to get_qat_config_config where it has full information about data type.
* Added support to TensorRT quantization
* Included check to account for when weight_qconfig_kwatgs is None.
* Removed output quantization from conv layers
* Added _Add_ReLU module that enables QATWrapper for quantizaiton.
* Removed quantization of output for linear and conv layers by default. Removed fusing of BN and ReLU by default.
* Minor fixes. Style and quality fixes.
* Added support to freezing bn stats.
* Added mode argument to wrapping of train function in BNWrapper
* Set BN fusing back as default.
* Set BN fusing back as default.
* Fixed custom freeze_bn_stats.
* Temporary files for evaluating changes to graphs.
* Added support to tensorrt flag. Moved the computation of quantization range to get_qat_config_config where it has full information about data type.
* Added support to TensorRT quantization
* Included check to account for when weight_qconfig_kwatgs is None.
* Modified argument names for backwards compatibility.
* Updated documentation to reflect changes.
* Updated documentation to reflect changes.
* Updated documentation to reflect changes.
* Fixed default weights data type.
* Style and quality fixes.
* Removed unused method
* Removed output quantization from conv layers
* Added _Add_ReLU module that enables QATWrapper for quantizaiton.
* Removed quantization of output for linear and conv layers by default. Removed fusing of BN and ReLU by default.
* Minor fixes. Style and quality fixes.
* Added support to freezing bn stats.
* Added mode argument to wrapping of train function in BNWrapper
* Set BN fusing back as default.
* Set BN fusing back as default.
* Fixed custom freeze_bn_stats.
* Temporary files for evaluating changes to graphs.
* Added support to tensorrt flag. Moved the computation of quantization range to get_qat_config_config where it has full information about data type.
* Added support to TensorRT quantization
* Included check to account for when weight_qconfig_kwatgs is None.
* Removed output quantization from conv layers
* Added _Add_ReLU module that enables QATWrapper for quantization.
* Removed quantization of output for linear and conv layers by default. Removed fusing of BN and ReLU by default.
* Minor fixes. Style and quality fixes.
* Added support to freezing bn stats.
* Added mode argument to wrapping of train function in BNWrapper
* Set BN fusing back as default.
* Set BN fusing back as default.
* Fixed custom freeze_bn_stats.
* Temporary files for evaluating changes to graphs.
* Added support to tensorrt flag. Moved the computation of quantization range to get_qat_config_config where it has full information about data type.
* Added support to TensorRT quantization
* Included check to account for when weight_qconfig_kwatgs is None.
* Modified argument names for backwards compatibility.
* Updated documentation to reflect changes.
* Updated documentation to reflect changes.
* Updated documentation to reflect changes.
* Fixed default weights data type.
* Style and quality fixes.
* Removed unused method
* Removed output quantization from conv layers
* Added _Add_ReLU module that enables QATWrapper for quantizaiton.
* Removed quantization of output for linear and conv layers by default. Removed fusing of BN and ReLU by default.
* Minor fixes. Style and quality fixes.
* Added support to freezing bn stats.
* Added mode argument to wrapping of train function in BNWrapper
* Set BN fusing back as default.
* Set BN fusing back as default.
* Fixed custom freeze_bn_stats.
* Temporary files for evaluating changes to graphs.
* Added support to tensorrt flag. Moved the computation of quantization range to get_qat_config_config where it has full information about data type.
* Added support to TensorRT quantization
* Included check to account for when weight_qconfig_kwatgs is None.
* Removed output quantization from conv layers
* Added _Add_ReLU module that enables QATWrapper for quantization.
* Removed quantization of output for linear and conv layers by default. Removed fusing of BN and ReLU by default.
* Minor fixes. Style and quality fixes.
* Added support to freezing bn stats.
* Added mode argument to wrapping of train function in BNWrapper
* Set BN fusing back as default.
* Set BN fusing back as default.
* Fixed custom freeze_bn_stats.
* Temporary files for evaluating changes to graphs.
* Added support to tensorrt flag. Moved the computation of quantization range to get_qat_config_config where it has full information about data type.
* Added support to TensorRT quantization
* Included check to account for when weight_qconfig_kwatgs is None.
* Modified argument names for backwards compatibility.
* Updated documentation to reflect changes.
* Updated documentation to reflect changes.
* Updated documentation to reflect changes.
* Fixed default weights data type.
* Style and quality fixes.
* Removed unused method
* Removed output quantization from conv layers
* Added _Add_ReLU module that enables QATWrapper for quantizaiton.
* Removed quantization of output for linear and conv layers by default. Removed fusing of BN and ReLU by default.
* Minor fixes. Style and quality fixes.
* Added support to freezing bn stats.
* Added mode argument to wrapping of train function in BNWrapper
* Set BN fusing back as default.
* Set BN fusing back as default.
* Fixed custom freeze_bn_stats.
* Temporary files for evaluating changes to graphs.
* Added support to tensorrt flag. Moved the computation of quantization range to get_qat_config_config where it has full information about data type.
* Added support to TensorRT quantization
* Included check to account for when weight_qconfig_kwatgs is None.
* Removed output quantization from conv layers
* Added _Add_ReLU module that enables QATWrapper for quantizaiton.
* Removed quantization of output for linear and conv layers by default. Removed fusing of BN and ReLU by default.
* Minor fixes. Style and quality fixes.
* Added support to freezing bn stats.
* Added mode argument to wrapping of train function in BNWrapper
* Set BN fusing back as default.
* Set BN fusing back as default.
* Fixed custom freeze_bn_stats.
* Temporary files for evaluating changes to graphs.
* Added support to tensorrt flag. Moved the computation of quantization range to get_qat_config_config where it has full information about data type.
* Added support to TensorRT quantization
* Included check to account for when weight_qconfig_kwatgs is None.
* Modified argument names for backwards compatibility.
* Updated documentation to reflect changes.
* Updated documentation to reflect changes.
* Updated documentation to reflect changes.
* Fixed default weights data type.
* Style and quality fixes.
* Removed unused method
* Removed testing files
* Style and quality fixes.
* Changed call to get_qat_qconfig to not specify symmetry and data type arguments for default case.
* Removed output quantization from conv layers
* Added _Add_ReLU module that enables QATWrapper for quantizaiton.
* Removed quantization of output for linear and conv layers by default. Removed fusing of BN and ReLU by default.
* Minor fixes. Style and quality fixes.
* Added support to freezing bn stats.
* Added mode argument to wrapping of train function in BNWrapper
* Set BN fusing back as default.
* Set BN fusing back as default.
* Fixed custom freeze_bn_stats.
* Temporary files for evaluating changes to graphs.
* Added support to tensorrt flag. Moved the computation of quantization range to get_qat_config_config where it has full information about data type.
* Added support to TensorRT quantization
* Included check to account for when weight_qconfig_kwatgs is None.
* Removed output quantization from conv layers
* Added _Add_ReLU module that enables QATWrapper for quantization.
* Removed quantization of output for linear and conv layers by default. Removed fusing of BN and ReLU by default.
* Minor fixes. Style and quality fixes.
* Added support to freezing bn stats.
* Added mode argument to wrapping of train function in BNWrapper
* Set BN fusing back as default.
* Set BN fusing back as default.
* Fixed custom freeze_bn_stats.
* Temporary files for evaluating changes to graphs.
* Added support to tensorrt flag. Moved the computation of quantization range to get_qat_config_config where it has full information about data type.
* Added support to TensorRT quantization
* Included check to account for when weight_qconfig_kwatgs is None.
* Modified argument names for backwards compatibility.
* Updated documentation to reflect changes.
* Updated documentation to reflect changes.
* Updated documentation to reflect changes.
* Fixed default weights data type.
* Style and quality fixes.
* Removed unused method
* Removed output quantization from conv layers
* Added _Add_ReLU module that enables QATWrapper for quantizaiton.
* Removed quantization of output for linear and conv layers by default. Removed fusing of BN and ReLU by default.
* Minor fixes. Style and quality fixes.
* Added support to freezing bn stats.
* Added mode argument to wrapping of train function in BNWrapper
* Set BN fusing back as default.
* Set BN fusing back as default.
* Fixed custom freeze_bn_stats.
* Temporary files for evaluating changes to graphs.
* Added support to tensorrt flag. Moved the computation of quantization range to get_qat_config_config where it has full information about data type.
* Added support to TensorRT quantization
* Included check to account for when weight_qconfig_kwatgs is None.
* Removed output quantization from conv layers
* Added _Add_ReLU module that enables QATWrapper for quantizaiton.
* Removed quantization of output for linear and conv layers by default. Removed fusing of BN and ReLU by default.
* Minor fixes. Style and quality fixes.
* Added support to freezing bn stats.
* Added mode argument to wrapping of train function in BNWrapper
* Set BN fusing back as default.
* Set BN fusing back as default.
* Fixed custom freeze_bn_stats.
* Temporary files for evaluating changes to graphs.
* Added support to tensorrt flag. Moved the computation of quantization range to get_qat_config_config where it has full information about data type.
* Added support to TensorRT quantization
* Included check to account for when weight_qconfig_kwatgs is None.
* Modified argument names for backwards compatibility.
* Updated documentation to reflect changes.
* Updated documentation to reflect changes.
* Updated documentation to reflect changes.
* Fixed default weights data type.
* Style and quality fixes.
* Removed unused method
* Removed testing files
* Style and quality fixes.
* Changed call to get_qat_qconfig to not specify symmetry and data type arguments for default case.
* Changed default number of activation and weight bits from None to 8.
* Revert "Changed default number of activation and weight bits from None to 8."
This reverts commit 95e966ed929fa3512331a73667d5ba2ac3d594b1.
* Revert "Changed call to get_qat_qconfig to not specify symmetry and data type arguments for default case."
This reverts commit a675813.
* Lumped qconfig properties into a dataclass.
* Lumped qconfig properties into a dataclass.
* Lumped qconfig properties into a dataclass.
* Resetting conv and linear activation flags to True.
* Renamed class BNWrapper as _BNWrapper.
* Added logging messages for when tensorrt forces overriding of configs.
* Style and quality fixes.
* ConvInteger quantization conversion for quant refactor (#644)
* ConvInteger quantization conversion for quant refactor
* [quantization-refactor] mark/propagate conv export mode (#672)
* batch norm fold with existing bias param bug fix
* Quantization Refactor Tests (#685)
* rebase import fix
* update manager serialization test cases for new quantization params
Co-authored-by: Benjamin Fineran <[email protected]>
Co-authored-by: spacemanidol <[email protected]>
Co-authored-by: Benjamin <[email protected]>1 parent a612e7b commit 1db70be
File tree
8 files changed
+794
-317
lines changed- src/sparseml/pytorch
- models/classification
- sparsification/quantization
- utils
- tests/sparseml/pytorch
- optim
- sparsification/quantization
8 files changed
+794
-317
lines changed| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
47 | 47 | | |
48 | 48 | | |
49 | 49 | | |
50 | | - | |
51 | 50 | | |
52 | 51 | | |
53 | 52 | | |
| |||
141 | 140 | | |
142 | 141 | | |
143 | 142 | | |
| 143 | + | |
| 144 | + | |
| 145 | + | |
| 146 | + | |
| 147 | + | |
| 148 | + | |
| 149 | + | |
| 150 | + | |
| 151 | + | |
| 152 | + | |
| 153 | + | |
| 154 | + | |
| 155 | + | |
| 156 | + | |
| 157 | + | |
| 158 | + | |
| 159 | + | |
| 160 | + | |
| 161 | + | |
| 162 | + | |
| 163 | + | |
| 164 | + | |
144 | 165 | | |
145 | 166 | | |
146 | 167 | | |
| |||
164 | 185 | | |
165 | 186 | | |
166 | 187 | | |
167 | | - | |
168 | | - | |
169 | | - | |
170 | | - | |
171 | | - | |
| 188 | + | |
172 | 189 | | |
173 | 190 | | |
174 | 191 | | |
| |||
181 | 198 | | |
182 | 199 | | |
183 | 200 | | |
184 | | - | |
185 | | - | |
186 | | - | |
187 | | - | |
188 | | - | |
189 | | - | |
| 201 | + | |
190 | 202 | | |
191 | 203 | | |
192 | 204 | | |
| |||
230 | 242 | | |
231 | 243 | | |
232 | 244 | | |
233 | | - | |
234 | | - | |
235 | | - | |
236 | | - | |
237 | | - | |
| 245 | + | |
238 | 246 | | |
239 | 247 | | |
240 | 248 | | |
| |||
252 | 260 | | |
253 | 261 | | |
254 | 262 | | |
255 | | - | |
256 | | - | |
257 | | - | |
258 | | - | |
259 | | - | |
| 263 | + | |
260 | 264 | | |
261 | 265 | | |
262 | 266 | | |
| |||
0 commit comments