Commit 1a8c5cc
Add AWQ mapping for GlmMoeDsaForCausalLM (GLM-5) (vllm-project#2418)
Closes vllm-project#2412 (part of vllm-project#1442)
GLM-5 (`GlmMoeDsaForCausalLM`) uses Multi-head Latent Attention
identical to DeepSeek V3 — same projection layer names (`q_a_proj`,
`kv_a_proj_with_mqa`, etc.) and same MoE structure. Reuses
`_deepseek_mappings`.
Also moves `Glm4MoeForCausalLM` to its correct alphabetical position in
the registry.
SUMMARY:
Add GLM-5 (GlmMoeDsaForCausalLM) to AWQ_MAPPING_REGISTRY using
_deepseek_mappings. GLM-5's MLA layer names are identical to DeepSeek
V3. Also fixes alphabetical ordering of Glm4MoeForCausalLM.
TEST PLAN:
Registry-only change (no logic modified). Verified that GLM-5 layer
names (q_a_proj, kv_a_proj_with_mqa, kv_a_layernorm, kv_b_proj, o_proj)
match the patterns in _deepseek_mappings by inspecting the
GlmMoeDsaForCausalLM source in transformers.
Signed-off-by: gillesturpin <turpingilles@orange.fr>
Co-authored-by: Brian Dellabetta <brian-dellabetta@users.noreply.github.com>
Co-authored-by: HDCharles <39544797+HDCharles@users.noreply.github.com>
Signed-off-by: yiliu30 <yi4.liu@intel.com>1 parent bbb634a commit 1a8c5cc
1 file changed
+2
-1
lines changed| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
251 | 251 | | |
252 | 252 | | |
253 | 253 | | |
| 254 | + | |
| 255 | + | |
254 | 256 | | |
255 | 257 | | |
256 | 258 | | |
| |||
264 | 266 | | |
265 | 267 | | |
266 | 268 | | |
267 | | - | |
268 | 269 | | |
269 | 270 | | |
270 | 271 | | |
| |||
0 commit comments