@@ -303,7 +303,7 @@ class TensorNameMap:
303303 "model.layers.{bid}.block_sparse_moe.router.layer" , # granitemoe
304304 "model.layers.{bid}.feed_forward.router" , # llama4
305305 "encoder.layers.{bid}.mlp.router.layer" , # nomic-bert-moe
306- "model.layers.{bid}.mlp.gate.wg.weight" , # hunyuan
306+ "model.layers.{bid}.mlp.gate.wg" , # hunyuan
307307 ),
308308
309309 MODEL_TENSOR .FFN_GATE_INP_SHEXP : (
@@ -363,7 +363,7 @@ class TensorNameMap:
363363 "model.layers.{bid}.mlp.shared_expert.up_proj" , # qwen2moe
364364 "model.layers.{bid}.mlp.shared_experts.up_proj" , # deepseek deepseek2
365365 "model.layers.{bid}.feed_forward.shared_expert.up_proj" , # llama4
366- "model.layers.{bid}.mlp.shared_mlp.up_proj.weight" , # hunyuan
366+ "model.layers.{bid}.mlp.shared_mlp.up_proj" , # hunyuan
367367 ),
368368
369369 # AWQ-activation gate
@@ -400,7 +400,7 @@ class TensorNameMap:
400400 "model.layers.{bid}.mlp.shared_expert.gate_proj" , # qwen2moe
401401 "model.layers.{bid}.mlp.shared_experts.gate_proj" , # deepseek deepseek2
402402 "model.layers.{bid}.feed_forward.shared_expert.gate_proj" , # llama4
403- "model.layers.{bid}.mlp.shared_mlp.gate_proj.weight" , # hunyuan
403+ "model.layers.{bid}.mlp.shared_mlp.gate_proj" , # hunyuan
404404 ),
405405
406406 # Feed-forward down
@@ -450,7 +450,7 @@ class TensorNameMap:
450450 "model.layers.{bid}.mlp.shared_experts.down_proj" , # deepseek deepseek2
451451 "model.layers.{bid}.feed_forward.shared_expert.down_proj" , # llama4
452452 "model.layers.{bid}.shared_mlp.output_linear" , # granitemoe
453- "model.layers.{bid}.mlp.shared_mlp.down_proj.weight" , # hunyuan
453+ "model.layers.{bid}.mlp.shared_mlp.down_proj" , # hunyuan
454454 ),
455455
456456 MODEL_TENSOR .ATTN_Q_NORM : (
0 commit comments