@@ -579,6 +579,10 @@ static const std::map<llm_arch, std::map<llm_tensor, const char *>> LLM_TENSOR_N
579579 { LLM_TENSOR_FFN_GATE, " blk.%d.ffn_gate" },
580580 { LLM_TENSOR_FFN_DOWN, " blk.%d.ffn_down" },
581581 { LLM_TENSOR_FFN_UP, " blk.%d.ffn_up" },
582+ { LLM_TENSOR_MTP_INP_PROJ, " blk.%d.mtp_inp_proj" },
583+ { LLM_TENSOR_MTP_TOKEN_NORM, " blk.%d.mtp_token_norm" },
584+ { LLM_TENSOR_MTP_HIDDEN_NORM, " blk.%d.mtp_hidden_norm" },
585+ { LLM_TENSOR_LAYER_OUT_NORM, " blk.%d.layer_output_norm" },
582586 },
583587 },
584588 {
@@ -1678,6 +1682,9 @@ static const std::map<llm_tensor, llm_tensor_info> LLM_TENSOR_INFOS = {
16781682 {LLM_TENSOR_FFN_GATE_EXPS, {LLM_TENSOR_LAYER_REPEATING, GGML_OP_MUL_MAT_ID}},
16791683 {LLM_TENSOR_FFN_UP_EXPS, {LLM_TENSOR_LAYER_REPEATING, GGML_OP_MUL_MAT_ID}},
16801684 {LLM_TENSOR_FFN_EXP_PROBS_B, {LLM_TENSOR_LAYER_REPEATING, GGML_OP_ADD}},
1685+ {LLM_TENSOR_MTP_INP_PROJ, {LLM_TENSOR_LAYER_REPEATING, GGML_OP_MUL_MAT}},
1686+ {LLM_TENSOR_MTP_TOKEN_NORM, {LLM_TENSOR_LAYER_REPEATING, GGML_OP_MUL}},
1687+ {LLM_TENSOR_MTP_HIDDEN_NORM, {LLM_TENSOR_LAYER_REPEATING, GGML_OP_MUL}},
16811688 // this tensor is loaded for T5, but never used
16821689 {LLM_TENSOR_DEC_CROSS_ATTN_REL_B, {LLM_TENSOR_LAYER_REPEATING, GGML_OP_NONE}},
16831690 {LLM_TENSOR_CONV1D, {LLM_TENSOR_LAYER_INPUT, GGML_OP_IM2COL}},
0 commit comments