@@ -90,6 +90,7 @@ static const std::map<llm_arch, const char *> LLM_ARCH_NAMES = {
9090 { LLM_ARCH_BAILINGMOE2, " bailingmoe2" },
9191 { LLM_ARCH_DOTS1, " dots1" },
9292 { LLM_ARCH_ARCEE, " arcee" },
93+ { LLM_ARCH_AFMOE, " afmoe" },
9394 { LLM_ARCH_ERNIE4_5, " ernie4_5" },
9495 { LLM_ARCH_ERNIE4_5_MOE, " ernie4_5-moe" },
9596 { LLM_ARCH_HUNYUAN_MOE, " hunyuan-moe" },
@@ -333,6 +334,36 @@ static const std::map<llm_arch, std::map<llm_tensor, const char *>> LLM_TENSOR_N
333334 { LLM_TENSOR_FFN_UP, " blk.%d.ffn_up" },
334335 },
335336 },
337+ {
338+ LLM_ARCH_AFMOE,
339+ {
340+ { LLM_TENSOR_TOKEN_EMBD, " token_embd" },
341+ { LLM_TENSOR_OUTPUT_NORM, " output_norm" },
342+ { LLM_TENSOR_OUTPUT, " output" },
343+ { LLM_TENSOR_ATTN_NORM, " blk.%d.attn_norm" },
344+ { LLM_TENSOR_ATTN_POST_NORM, " blk.%d.post_attention_norm" },
345+ { LLM_TENSOR_ATTN_Q, " blk.%d.attn_q" },
346+ { LLM_TENSOR_ATTN_K, " blk.%d.attn_k" },
347+ { LLM_TENSOR_ATTN_V, " blk.%d.attn_v" },
348+ { LLM_TENSOR_ATTN_OUT, " blk.%d.attn_output" },
349+ { LLM_TENSOR_ATTN_Q_NORM, " blk.%d.attn_q_norm" },
350+ { LLM_TENSOR_ATTN_K_NORM, " blk.%d.attn_k_norm" },
351+ { LLM_TENSOR_ATTN_GATE, " blk.%d.attn_gate" },
352+ { LLM_TENSOR_FFN_NORM, " blk.%d.ffn_norm" },
353+ { LLM_TENSOR_FFN_POST_NORM, " blk.%d.post_ffw_norm" },
354+ { LLM_TENSOR_FFN_GATE_INP, " blk.%d.ffn_gate_inp" },
355+ { LLM_TENSOR_FFN_GATE, " blk.%d.ffn_gate" },
356+ { LLM_TENSOR_FFN_DOWN, " blk.%d.ffn_down" },
357+ { LLM_TENSOR_FFN_UP, " blk.%d.ffn_up" },
358+ { LLM_TENSOR_FFN_GATE_EXPS, " blk.%d.ffn_gate_exps" },
359+ { LLM_TENSOR_FFN_DOWN_EXPS, " blk.%d.ffn_down_exps" },
360+ { LLM_TENSOR_FFN_UP_EXPS, " blk.%d.ffn_up_exps" },
361+ { LLM_TENSOR_FFN_GATE_SHEXP, " blk.%d.ffn_gate_shexp" },
362+ { LLM_TENSOR_FFN_UP_SHEXP, " blk.%d.ffn_up_shexp" },
363+ { LLM_TENSOR_FFN_DOWN_SHEXP, " blk.%d.ffn_down_shexp" },
364+ { LLM_TENSOR_FFN_EXP_PROBS_B, " blk.%d.exp_probs_b" },
365+ },
366+ },
336367 {
337368 LLM_ARCH_LLAMA4,
338369 {
@@ -2444,6 +2475,7 @@ static const std::map<llm_tensor, llm_tensor_info> LLM_TENSOR_INFOS = {
24442475 {LLM_TENSOR_ATTN_V, {LLM_TENSOR_LAYER_REPEATING, GGML_OP_MUL_MAT}},
24452476 {LLM_TENSOR_ATTN_QKV, {LLM_TENSOR_LAYER_REPEATING, GGML_OP_MUL_MAT}},
24462477 {LLM_TENSOR_ATTN_OUT, {LLM_TENSOR_LAYER_REPEATING, GGML_OP_MUL_MAT}},
2478+ {LLM_TENSOR_ATTN_GATE, {LLM_TENSOR_LAYER_REPEATING, GGML_OP_MUL_MAT}},
24472479 {LLM_TENSOR_FFN_GATE, {LLM_TENSOR_LAYER_REPEATING, GGML_OP_MUL_MAT}},
24482480 {LLM_TENSOR_FFN_DOWN, {LLM_TENSOR_LAYER_REPEATING, GGML_OP_MUL_MAT}},
24492481 {LLM_TENSOR_FFN_UP, {LLM_TENSOR_LAYER_REPEATING, GGML_OP_MUL_MAT}},
0 commit comments