Skip to content

Commit c53d4dc

Browse files
committed
Add AWQ mapping for GlmMoeDsaForCausalLM (GLM-5)
GLM-5 uses Multi-head Latent Attention identical to DeepSeek V3 — same projection names (q_a_proj, kv_a_proj_with_mqa, etc.) and same MoE structure. Reuses _deepseek_mappings. Also moves Glm4MoeForCausalLM to its correct alphabetical position.
1 parent a9847e0 commit c53d4dc

File tree

1 file changed

+2
-1
lines changed

1 file changed

+2
-1
lines changed

src/llmcompressor/modifiers/awq/mappings.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -251,6 +251,8 @@ class AWQMapping:
251251
"Gemma2ForCausalLM": _gemma_mappings,
252252
"Gemma3ForCausalLM": _gemma_mappings,
253253
"Gemma3ForConditionalGeneration": _gemma_mappings,
254+
"Glm4MoeForCausalLM": _default_mappings,
255+
"GlmMoeDsaForCausalLM": _deepseek_mappings,
254256
"LlamaForCausalLM": _default_mappings,
255257
"Llama4ForConditionalGeneration": _default_mappings,
256258
"Mistral3ForConditionalGeneration": _default_mappings,
@@ -264,7 +266,6 @@ class AWQMapping:
264266
"Qwen3ForCausalLM": _default_mappings,
265267
"Qwen3MoeForCausalLM": _moe_default_mappings,
266268
"Qwen3NextForCausalLM": _qwen3_next_moe_mappings,
267-
"Glm4MoeForCausalLM": _default_mappings,
268269
"SeedOssForCausalLM": _default_mappings,
269270
"Ernie4_5_MoeForCausalLM": _default_mappings,
270271
}

0 commit comments

Comments
 (0)