@@ -453,9 +453,6 @@ class MODEL_TENSOR(IntEnum):
453453 FFN_GATE_EXP = auto ()
454454 FFN_DOWN_EXP = auto ()
455455 FFN_UP_EXP = auto ()
456- FFN_GATE_EXPS = auto ()
457- FFN_DOWN_EXPS = auto ()
458- FFN_UP_EXPS = auto ()
459456 FFN_GATE_SHEXP = auto ()
460457 FFN_DOWN_SHEXP = auto ()
461458 FFN_UP_SHEXP = auto ()
@@ -807,9 +804,6 @@ class MODEL_TENSOR(IntEnum):
807804 MODEL_TENSOR .FFN_GATE_EXP : "blk.{bid}.ffn_gate_exps" ,
808805 MODEL_TENSOR .FFN_DOWN_EXP : "blk.{bid}.ffn_down_exps" ,
809806 MODEL_TENSOR .FFN_UP_EXP : "blk.{bid}.ffn_up_exps" ,
810- MODEL_TENSOR .FFN_GATE_EXPS : "blk.{bid}.ffn_gate_exps" ,
811- MODEL_TENSOR .FFN_DOWN_EXPS : "blk.{bid}.ffn_down_exps" ,
812- MODEL_TENSOR .FFN_UP_EXPS : "blk.{bid}.ffn_up_exps" ,
813807 MODEL_TENSOR .FFN_EXP_PROBS_B : "blk.{bid}.exp_probs_b" ,
814808 MODEL_TENSOR .LAYER_OUT_NORM : "blk.{bid}.layer_output_norm" ,
815809 MODEL_TENSOR .PER_LAYER_TOKEN_EMBD : "per_layer_token_embd" , # gemma3n
@@ -2581,9 +2575,9 @@ class MODEL_TENSOR(IntEnum):
25812575 MODEL_TENSOR .FFN_DOWN ,
25822576 MODEL_TENSOR .FFN_UP ,
25832577 MODEL_TENSOR .FFN_GATE_INP ,
2584- MODEL_TENSOR .FFN_GATE_EXPS ,
2585- MODEL_TENSOR .FFN_DOWN_EXPS ,
2586- MODEL_TENSOR .FFN_UP_EXPS ,
2578+ MODEL_TENSOR .FFN_GATE_EXP ,
2579+ MODEL_TENSOR .FFN_DOWN_EXP ,
2580+ MODEL_TENSOR .FFN_UP_EXP ,
25872581 MODEL_TENSOR .FFN_GATE_SHEXP ,
25882582 MODEL_TENSOR .FFN_UP_SHEXP ,
25892583 MODEL_TENSOR .FFN_DOWN_SHEXP ,
0 commit comments