Skip to content

Fixed FP8 fake quantization to use fp32 amax scaling; Added support for FP8 per-channel quantization#381

Merged
realAsma merged 2 commits intomainfrom
asma/fp8_fakequant_fix
Sep 30, 2025
Merged

Fixed FP8 fake quantization to use fp32 amax scaling; Added support for FP8 per-channel quantization#381
realAsma merged 2 commits intomainfrom
asma/fp8_fakequant_fix

Commits

Commits on Sep 30, 2025