@@ -2237,7 +2237,7 @@ static const std::map<llm_arch, std::map<llm_tensor, const char *>> LLM_TENSOR_N
22372237 { LLM_TENSOR_SHORTCONV_INPROJ, " blk.%d.shortconv.in_proj" },
22382238 { LLM_TENSOR_SHORTCONV_OUTPROJ, " blk.%d.shortconv.out_proj" },
22392239 { LLM_TENSOR_TOKEN_EMBD, " token_embd" },
2240- { LLM_TENSOR_TOKEN_EMBD_NORM , " token_embd_norm" },
2240+ { LLM_TENSOR_OUTPUT_NORM , " token_embd_norm" }, // note: wrong tensor name
22412241 { LLM_TENSOR_OUTPUT, " output" },
22422242 }
22432243 },
@@ -2259,7 +2259,7 @@ static const std::map<llm_arch, std::map<llm_tensor, const char *>> LLM_TENSOR_N
22592259 { LLM_TENSOR_SHORTCONV_INPROJ, " blk.%d.shortconv.in_proj" },
22602260 { LLM_TENSOR_SHORTCONV_OUTPROJ, " blk.%d.shortconv.out_proj" },
22612261 { LLM_TENSOR_TOKEN_EMBD, " token_embd" },
2262- { LLM_TENSOR_TOKEN_EMBD_NORM , " token_embd_norm" },
2262+ { LLM_TENSOR_OUTPUT_NORM , " token_embd_norm" }, // note: wrong tensor name
22632263 { LLM_TENSOR_FFN_GATE_INP, " blk.%d.ffn_gate_inp" },
22642264 { LLM_TENSOR_FFN_GATE_EXPS, " blk.%d.ffn_gate_exps" },
22652265 { LLM_TENSOR_FFN_DOWN_EXPS, " blk.%d.ffn_down_exps" },
@@ -2490,8 +2490,8 @@ static const std::map<llm_arch, std::map<llm_tensor, const char *>> LLM_TENSOR_N
24902490static const std::map<llm_tensor, llm_tensor_info> LLM_TENSOR_INFOS = {
24912491 {LLM_TENSOR_TOKEN_EMBD, {LLM_TENSOR_LAYER_INPUT, GGML_OP_GET_ROWS}},
24922492 {LLM_TENSOR_POS_EMBD, {LLM_TENSOR_LAYER_INPUT, GGML_OP_GET_ROWS}},
2493- {LLM_TENSOR_TOKEN_EMBD_NORM, {LLM_TENSOR_LAYER_INPUT, GGML_OP_GET_ROWS}},
24942493 {LLM_TENSOR_TOKEN_TYPES, {LLM_TENSOR_LAYER_INPUT, GGML_OP_GET_ROWS}},
2494+ {LLM_TENSOR_TOKEN_EMBD_NORM, {LLM_TENSOR_LAYER_INPUT, GGML_OP_MUL}},
24952495 {LLM_TENSOR_OUTPUT, {LLM_TENSOR_LAYER_OUTPUT, GGML_OP_MUL_MAT}},
24962496 {LLM_TENSOR_CLS, {LLM_TENSOR_LAYER_OUTPUT, GGML_OP_MUL_MAT}},
24972497 {LLM_TENSOR_CLS_OUT, {LLM_TENSOR_LAYER_OUTPUT, GGML_OP_MUL_MAT}},
0 commit comments