Skip to content

Commit 7ae15d9

Browse files
authored
add nvfp4 benchmark (Tencent#77)
1 parent 6adea79 commit 7ae15d9

File tree

3 files changed

+25
-0
lines changed

3 files changed

+25
-0
lines changed

README.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -31,6 +31,7 @@
3131
- [技术交流](#技术交流)
3232

3333
## 📣最新进展
34+
- [25/09/24] 我们支持了Qwen3系列模型的NVFP4的PTQ量化,我们还开源了[Qwen3-32B-NVFP4](https://huggingface.co/AngelSlim/Qwen3-32B_nvfp4)[Qwen3-235B-A22B-NVFP4](https://huggingface.co/AngelSlim/Qwen3-235B-A22B_nvfp4)权重。
3435
- [25/09/01] 我们支持了[Hunyuan-MT-7B](https://huggingface.co/tencent/Hunyuan-MT-7B-fp8)翻译开源模型的FP8量化;支持了Eagle3的Torch推理及Benchmark评测流程;支持了[FLUX](https://github.com/Tencent/AngelSlim/tree/main/configs/flux)的量化、Cache;支持了[Seed-OSS](https://github.com/Tencent/AngelSlim/tree/main/configs/seed_oss)模型量化压缩。
3536
- [25/08/06] 我们支持了`Hunyuan 0.5B/1.8B/4B/7B``Qwen2.5VL 3B/7B/32B/72B`的FP8、INT4量化,支持了`DeepSeek-R1/V3``Kimi-K2`模型的`FP8-Static``W4A8-FP8`量化。我们还开源了`Hunyuan 1.8B/4B/7B`系列模型的Eagle3权重。
3637
- [25/07/04] 我们支持了`Hunyuan/Qwen2.5/Qwen3/DeepSeek-R1-Distill-Qwen`等模型的量化,包含INT8、FP8、INT4等算法。

README_en.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -31,6 +31,7 @@ Dedicated to building a more intuitive, comprehensive, and efficient LLMs compre
3131
- [Technical Discussion](#technical-discussion)
3232

3333
## 📣Latest Updates
34+
- [25/09/24] We now support the PTQ quantification of NVFP4 for the Qwen3 series models. We also opensource [Qwen3-32B-NVFP4](https://huggingface.co/AngelSlim/Qwen3-32B_nvfp4) and [Qwen3-235B-A22B-NVFP4](https://huggingface.co/AngelSlim/Qwen3-235B-A22B_nvfp4) weights.
3435
- [25/09/01] We now support ​FP8 quantization​ of the [Hunyuan-MT-7B](https://huggingface.co/tencent/Hunyuan-MT-7B-fp8) translation model. And enabled ​Torch inference and Benchmark evaluation​ for Eagle3. And implemented support for ​quantization and Cache​ for [FLUX](https://github.com/Tencent/AngelSlim/tree/main/configs/flux). And support ​quantization​ for the [Seed-OSS](https://github.com/Tencent/AngelSlim/tree/main/configs/seed_oss).
3536
- [25/08/06] We now support quantization for `Hunyuan 0.5B/1.8B/4B/7B` and multimodal model `Qwen2.5VL 3B/7B/32B/72B`, including `FP8/INT4` algorithms, and quantization for `DeepSeek-R1/V3` and `Kimi-K2`, including `FP8-Static` and `W4A8-FP8` algorithms. We also opensource `Hunyuan 1.8B/4B/7B` series Eagle3 model weight.
3637
- [25/07/04] We now support quantization for `Hunyuan/Qwen2.5/Qwen3/DeepSeek-R1-Distill-Qwen` and other models, including `INT8/FP8/INT4` algorithms. We also opensource `Qwen3` series Eagle3 model weight.

docs/source/performance/quantization/benchmarks.md

Lines changed: 23 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -386,3 +386,26 @@ INT4-GPTAQ在`GSM8K`、`HUMANEVAL`、`GPQA Diamond`上的评测结果如下:
386386
| | INT4-GPTAQ | 69.52 | 37.20 | - |
387387
+-----------+--------------+-------+-----------+--------------+
388388
```
389+
390+
391+
## NVFP4
392+
393+
NVFP4在`GSM8K``MMLU``GPQA Diamond`上的评测结果如下:
394+
395+
```{eval-rst}
396+
.. table::
397+
:align: center
398+
:name: table-NVFP4-performance
399+
400+
+-----------------+--------------+-------+-------+--------------+
401+
| Model | Quantization | GSM8K | MMLU | GPQA Diamond |
402+
+=================+==============+=======+=======+==============+
403+
| Qwen3-32B | BF16 | 67.06 | 81.72 | 54.04 |
404+
+ +--------------+-------+-------+--------------+
405+
| | NVFP4 | 69.87 | 80.74 | 56.06 |
406+
+-----------------+--------------+-------+-------+--------------+
407+
| Qwen3-235B-A22B | BF16 | 96.63 | 62.73 | 60.60 |
408+
+ +--------------+-------+-------+--------------+
409+
| | NVFP4 | 96.17 | 62.09 | 60.10 |
410+
+-----------------+--------------+-------+-------+--------------+
411+
```

0 commit comments

Comments
 (0)