Compressed Tensors v0.7.0
What's Changed
- Make INT8 activation PRESET_SCHEMES explicit by @mgoin in #158
- Write the current version into model configs by @mgoin in #160
- [KV-Cache] Make k_scale, v_scale as attributes of self_attn using HFCache by @horheynm in #148
- [Bugfix] Fix quant config parsing by @kylesayrs in #162
- Ignore Dense sparsity config by @rahul-tuli in #169
- fix bug by @horheynm in #170
- Replace
compression_config
to bequantization_config
forHFQuantizer
support by @dsikka in #164 - ignore list by @horheynm in #171
- switch default to release and disable pushing to pypi for now by @dhuangnm in #175
- Fix missing quant_method value by @kylesayrs in #174
- Fix ModelCompressor parsing in HF Quantizer case by @kylesayrs in #176
- Calibration Code Clarity by @kylesayrs in #168
- Add: base sparsity/quantization compressors by @rahul-tuli in #165
- Update compressors folder structure by @rahul-tuli in #166
- Update number of groups by @dsikka in #178
- Bring nightly build/test back by @dhuangnm in #179
- Remove unused function by @kylesayrs in #156
- Revert "Ignore Dense sparsity config (#169)" by @rahul-tuli in #181
- Workaround HF Quantizer
apply_quantization_config
misuse by @kylesayrs in #180 - bump up version to 0.7.0 by @dhuangnm in #186
Full Changelog: 0.6.0...0.7.0