Skip to content

Commit 83244a9

Browse files
committed
Remove accidentally duplicated tensor constants
1 parent ae8364c commit 83244a9

File tree

2 files changed

+6
-12
lines changed

2 files changed

+6
-12
lines changed

convert_hf_to_gguf.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -2313,13 +2313,13 @@ def modify_tensors(self, data_torch: Tensor, name: str, bid: int | None) -> Iter
23132313
# Handle expert weights - they're already merged in the HF format
23142314
if ".block_sparse_moe.experts.w1" in name:
23152315
assert bid is not None
2316-
return [(self.format_tensor_name(gguf.MODEL_TENSOR.FFN_GATE_EXPS, bid), data_torch)]
2316+
return [(self.format_tensor_name(gguf.MODEL_TENSOR.FFN_GATE_EXP, bid), data_torch)]
23172317
elif ".block_sparse_moe.experts.w2" in name:
23182318
assert bid is not None
2319-
return [(self.format_tensor_name(gguf.MODEL_TENSOR.FFN_DOWN_EXPS, bid), data_torch)]
2319+
return [(self.format_tensor_name(gguf.MODEL_TENSOR.FFN_DOWN_EXP, bid), data_torch)]
23202320
elif ".block_sparse_moe.experts.w3" in name:
23212321
assert bid is not None
2322-
return [(self.format_tensor_name(gguf.MODEL_TENSOR.FFN_UP_EXPS, bid), data_torch)]
2322+
return [(self.format_tensor_name(gguf.MODEL_TENSOR.FFN_UP_EXP, bid), data_torch)]
23232323

23242324
# Map dual normalization layers
23252325
if ".attn_norm_a." in name and bid is not None:

gguf-py/gguf/constants.py

Lines changed: 3 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -453,9 +453,6 @@ class MODEL_TENSOR(IntEnum):
453453
FFN_GATE_EXP = auto()
454454
FFN_DOWN_EXP = auto()
455455
FFN_UP_EXP = auto()
456-
FFN_GATE_EXPS = auto()
457-
FFN_DOWN_EXPS = auto()
458-
FFN_UP_EXPS = auto()
459456
FFN_GATE_SHEXP = auto()
460457
FFN_DOWN_SHEXP = auto()
461458
FFN_UP_SHEXP = auto()
@@ -807,9 +804,6 @@ class MODEL_TENSOR(IntEnum):
807804
MODEL_TENSOR.FFN_GATE_EXP: "blk.{bid}.ffn_gate_exps",
808805
MODEL_TENSOR.FFN_DOWN_EXP: "blk.{bid}.ffn_down_exps",
809806
MODEL_TENSOR.FFN_UP_EXP: "blk.{bid}.ffn_up_exps",
810-
MODEL_TENSOR.FFN_GATE_EXPS: "blk.{bid}.ffn_gate_exps",
811-
MODEL_TENSOR.FFN_DOWN_EXPS: "blk.{bid}.ffn_down_exps",
812-
MODEL_TENSOR.FFN_UP_EXPS: "blk.{bid}.ffn_up_exps",
813807
MODEL_TENSOR.FFN_EXP_PROBS_B: "blk.{bid}.exp_probs_b",
814808
MODEL_TENSOR.LAYER_OUT_NORM: "blk.{bid}.layer_output_norm",
815809
MODEL_TENSOR.PER_LAYER_TOKEN_EMBD: "per_layer_token_embd", # gemma3n
@@ -2581,9 +2575,9 @@ class MODEL_TENSOR(IntEnum):
25812575
MODEL_TENSOR.FFN_DOWN,
25822576
MODEL_TENSOR.FFN_UP,
25832577
MODEL_TENSOR.FFN_GATE_INP,
2584-
MODEL_TENSOR.FFN_GATE_EXPS,
2585-
MODEL_TENSOR.FFN_DOWN_EXPS,
2586-
MODEL_TENSOR.FFN_UP_EXPS,
2578+
MODEL_TENSOR.FFN_GATE_EXP,
2579+
MODEL_TENSOR.FFN_DOWN_EXP,
2580+
MODEL_TENSOR.FFN_UP_EXP,
25872581
MODEL_TENSOR.FFN_GATE_SHEXP,
25882582
MODEL_TENSOR.FFN_UP_SHEXP,
25892583
MODEL_TENSOR.FFN_DOWN_SHEXP,

0 commit comments

Comments
 (0)