Skip to content

Regarding the Quantization Scheme for the Qwen-Next 80B Model W8A8 #2202

@vonchenplus

Description

@vonchenplus

Hi there,

I’m planning to quantize the Qwen-Next 80B model to W8A8 using a combination of SmoothQuant and GPTQ. However, I noticed that SmoothQuant does not currently support this model.

In this situation, would it be acceptable to use the default mapping method for SmoothQuant? If not, are there any recommended alternatives or best practices for achieving stable and accurate W8A8 quantization on this model?

Thanks in advance for your help!

Metadata

Metadata

Assignees

Labels

enhancementNew feature or requestqwenFor any PR / issue related to Qwen supportsmoothquantFor any issue / PR related to SmoothQuant support

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions