LLM 自适应 Shift-SmoothQuant #1894
Unanswered
egbertngu66
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
用自适应 Shift-SmoothQuant量化LLM,能达到W8A8(per-tensor)可用吗?https://github.com/PaddlePaddle/PaddleSlim/blob/develop/docs/zh_cn/tutorials/quant/advanced_quantization.md 表中的A8是per-tensor量化还是per-token量化?有相应的实验配置吗
Beta Was this translation helpful? Give feedback.
All reactions