Skip to content

Compressed Tensors v0.10.0

Compare
Choose a tag to compare
@dhuangnm dhuangnm released this 05 Jun 17:51
· 100 commits to main since this release
d7ce8ec

What's Changed

  • Updates to build system by @dbarbuzzi in #304
  • [Utils] add align_modules by @kylesayrs in #282
  • Enable module state_dict compression, simplify compression logic by @kylesayrs in #302
  • Fix _initialize_scale_zero_point initializing on the wrong device by @mgoin in #295
  • Revert "Enable module state_dict compression, simplify compression lo… by @kylesayrs in #306
  • [Bugfix] Fix shape calculation for group quantization by @kylesayrs in #308
  • Enable module state_dict compression, simplify compression logic by @kylesayrs in #307
  • Clarify decompression return type by @kylesayrs in #310
  • Clarify match_param_name return type by @kylesayrs in #312
  • [Compressor][NVFP4] Support FP4 Compression by @dsikka in #311
  • [NVFP4] Update FloatArgs and NVFP4 by @dsikka in #313
  • fix signatures on model_validator functions by @brian-dellabetta in #314
  • [Performance] Add memory compression and decompression pathways by @kylesayrs in #301
  • Model Compression: Set compression status by @kylesayrs in #318
  • [NVFP4] Enable Fp4 Quantization; introduce / apply global_scales by @dsikka in #315
  • [NVFP4] Skip fused global scale calculation if already fused by @dsikka in #322
  • Update default observer to be MSE by @shanjiaz in #300
  • [Misc] Generics typehinting for RegistryMixin by @kylesayrs in #320
  • Revert "Update default observer to be MSE (#300)" by @dsikka in #323
  • [NVFP4] Add tensor_group strategy; enable NVFP4 Activations by @dsikka in #317
  • [Transforms] Transform Args, Scheme, and Config by @kylesayrs in #321
  • [NVFP4] Expand dynamic types, clean-up conditions by @dsikka in #325
  • Use different runner for UPLOAD job by @dbarbuzzi in #327
  • [NVFP4] Use torch.compile when rounding to NVFP4 by @dsikka in #331
  • [Tests] Update test_fp8_quant.py by @dsikka in #337
  • [Tests] Fix test scale init for group quant by @dsikka in #338
  • [Quantization] Update group quantization by @dsikka in #336
  • [NVFP4] update global scale generation by @dsikka in #339
  • [Transform] Accelerate Utilities by @kylesayrs in #328
  • Model Compression: Delete offload by @kylesayrs in #319
  • [Decompression] Keep unused parameters when decompressing from memory by @kylesayrs in #340
  • [NVFP4] Small Nits by @dsikka in #341

New Contributors

Full Changelog: 0.9.4...0.10.0