v0.4.0
What's Changed
- feat: add guards to sawb recomputation by @andrea-fasoli in #131
- build: Move torchvision to an optional dependency by @BrandonGroth in #144
- fix: feat: fix for new transformers (>4.48) and new QLinear for INT8 training with HW emulation by @chichun-charlie-liu in #141
- chore(deps): Update transformers requirement from <4.52,>=4.45 to >=4.45,<4.53 by @dependabot[bot] in #127
- build: Move triton to an optional dependency by @BrandonGroth in #146
- chore(deps): Update accelerate requirement from !=0.34,<1.7,>=0.20.3 to >=0.20.3,!=0.34,<1.9 by @dependabot[bot] in #143
- build: Make non-essential dependencies optional by @BrandonGroth in #147
- fix: fix available_packages by @chichun-charlie-liu in #153
- fix: Saved qconfig recipe being overwritten with defaults by @BrandonGroth in #152
- fix: Remove gptqmodel Warning on startup by @BrandonGroth in #156
- fix: Remove llmcompressor oneshot import deprecation warning by @BrandonGroth in #157
- feat: addons for FP8 attention bmm, paged attention, and linear in FMS by @ani300 in #154
- feat: addons for FP8 attention bmm and linear in FMS by @andrea-fasoli in #149
- feat: add QA and MaskedLM task for FP8 encoder instantiation by @andrea-fasoli in #148
- feat: AIU sim for FP8 (DL8/DL16) added to triton kernel by @chichun-charlie-liu in #159
- fix: qkvsync bug fix by @chichun-charlie-liu in #161
- chore(deps): Update datasets requirement from <4.0,>=3.0.0 to >=3.0.0,<5.0 by @dependabot[bot] in #160
New Contributors
Full Changelog: v0.3.0...v0.4.0