@@ -1777,26 +1777,26 @@ static const std::map<llm_arch, std::map<llm_tensor, const char *>> LLM_TENSOR_N
17771777 },
17781778 },
17791779 {
1780- LLM_ARCH_UNKNOWN ,
1780+ LLM_ARCH_SMOLLM3 ,
17811781 {
1782- { LLM_TENSOR_TOKEN_EMBD, " token_embd" },
1782+ { LLM_TENSOR_TOKEN_EMBD, " token_embd" },
1783+ { LLM_TENSOR_OUTPUT_NORM, " output_norm" },
1784+ { LLM_TENSOR_OUTPUT, " output" },
1785+ { LLM_TENSOR_ATTN_NORM, " blk.%d.attn_norm" },
1786+ { LLM_TENSOR_ATTN_Q, " blk.%d.attn_q" },
1787+ { LLM_TENSOR_ATTN_K, " blk.%d.attn_k" },
1788+ { LLM_TENSOR_ATTN_V, " blk.%d.attn_v" },
1789+ { LLM_TENSOR_ATTN_OUT, " blk.%d.attn_output" },
1790+ { LLM_TENSOR_FFN_NORM, " blk.%d.ffn_norm" },
1791+ { LLM_TENSOR_FFN_GATE, " blk.%d.ffn_gate" },
1792+ { LLM_TENSOR_FFN_DOWN, " blk.%d.ffn_down" },
1793+ { LLM_TENSOR_FFN_UP, " blk.%d.ffn_up" },
17831794 },
17841795 },
17851796 {
1786- LLM_ARCH_SMOLLM3 ,
1797+ LLM_ARCH_UNKNOWN ,
17871798 {
1788- { LLM_TENSOR_TOKEN_EMBD, " token_embd" },
1789- { LLM_TENSOR_OUTPUT_NORM, " output_norm" },
1790- { LLM_TENSOR_OUTPUT, " output" },
1791- { LLM_TENSOR_ATTN_NORM, " blk.%d.attn_norm" },
1792- { LLM_TENSOR_ATTN_Q, " blk.%d.attn_q" },
1793- { LLM_TENSOR_ATTN_K, " blk.%d.attn_k" },
1794- { LLM_TENSOR_ATTN_V, " blk.%d.attn_v" },
1795- { LLM_TENSOR_ATTN_OUT, " blk.%d.attn_output" },
1796- { LLM_TENSOR_FFN_NORM, " blk.%d.ffn_norm" },
1797- { LLM_TENSOR_FFN_GATE, " blk.%d.ffn_gate" },
1798- { LLM_TENSOR_FFN_DOWN, " blk.%d.ffn_down" },
1799- { LLM_TENSOR_FFN_UP, " blk.%d.ffn_up" },
1799+ { LLM_TENSOR_TOKEN_EMBD, " token_embd" },
18001800 },
18011801 },
18021802};
0 commit comments