@@ -72,6 +72,7 @@ static const std::map<llm_arch, const char *> LLM_ARCH_NAMES = {
72
72
{ LLM_ARCH_WAVTOKENIZER_DEC, " wavtokenizer-dec" },
73
73
{ LLM_ARCH_PLM, " plm" },
74
74
{ LLM_ARCH_BAILINGMOE, " bailingmoe" },
75
+ { LLM_ARCH_DOTS1, " dots1" },
75
76
{ LLM_ARCH_UNKNOWN, " (unknown)" },
76
77
};
77
78
@@ -1555,6 +1556,34 @@ static const std::map<llm_arch, std::map<llm_tensor, const char *>> LLM_TENSOR_N
1555
1556
{ LLM_TENSOR_FFN_UP_SHEXP, " blk.%d.ffn_up_shexp" },
1556
1557
},
1557
1558
},
1559
+ {
1560
+ LLM_ARCH_DOTS1,
1561
+ {
1562
+ { LLM_TENSOR_TOKEN_EMBD, " token_embd" },
1563
+ { LLM_TENSOR_OUTPUT_NORM, " output_norm" },
1564
+ { LLM_TENSOR_OUTPUT, " output" },
1565
+ { LLM_TENSOR_ATTN_NORM, " blk.%d.attn_norm" },
1566
+ { LLM_TENSOR_ATTN_Q, " blk.%d.attn_q" },
1567
+ { LLM_TENSOR_ATTN_Q_NORM, " blk.%d.attn_q_norm" },
1568
+ { LLM_TENSOR_ATTN_K, " blk.%d.attn_k" },
1569
+ { LLM_TENSOR_ATTN_K_NORM, " blk.%d.attn_k_norm" },
1570
+ { LLM_TENSOR_ATTN_V, " blk.%d.attn_v" },
1571
+ { LLM_TENSOR_ATTN_OUT, " blk.%d.attn_output" },
1572
+ { LLM_TENSOR_FFN_NORM, " blk.%d.ffn_norm" },
1573
+ { LLM_TENSOR_FFN_GATE, " blk.%d.ffn_gate" },
1574
+ { LLM_TENSOR_FFN_UP, " blk.%d.ffn_up" },
1575
+ { LLM_TENSOR_FFN_DOWN, " blk.%d.ffn_down" },
1576
+ { LLM_TENSOR_FFN_GATE_INP, " blk.%d.ffn_gate_inp" },
1577
+ { LLM_TENSOR_FFN_GATE_EXPS, " blk.%d.ffn_gate_exps" },
1578
+ { LLM_TENSOR_FFN_DOWN_EXPS, " blk.%d.ffn_down_exps" },
1579
+ { LLM_TENSOR_FFN_UP_EXPS, " blk.%d.ffn_up_exps" },
1580
+ { LLM_TENSOR_FFN_GATE_INP_SHEXP, " blk.%d.ffn_gate_inp_shexp" },
1581
+ { LLM_TENSOR_FFN_GATE_SHEXP, " blk.%d.ffn_gate_shexp" },
1582
+ { LLM_TENSOR_FFN_DOWN_SHEXP, " blk.%d.ffn_down_shexp" },
1583
+ { LLM_TENSOR_FFN_UP_SHEXP, " blk.%d.ffn_up_shexp" },
1584
+ { LLM_TENSOR_FFN_EXP_PROBS_B, " blk.%d.exp_probs_b" },
1585
+ }
1586
+ },
1558
1587
{
1559
1588
LLM_ARCH_UNKNOWN,
1560
1589
{
0 commit comments