Skip to content

Commit a8ac707

Browse files
committed
add support for bitnet2b_2501 model
1 parent 957b59d commit a8ac707

File tree

3 files changed

+327
-1
lines changed

3 files changed

+327
-1
lines changed

gguf-py/gguf/constants.py

Lines changed: 24 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -237,6 +237,7 @@ class MODEL_ARCH(IntEnum):
237237
DEEPSEEK2 = auto()
238238
CHATGLM = auto()
239239
BITNET = auto()
240+
BITNET_25 = auto()
240241
T5 = auto()
241242
T5ENCODER = auto()
242243
JAIS = auto()
@@ -398,6 +399,7 @@ class MODEL_TENSOR(IntEnum):
398399
MODEL_ARCH.DEEPSEEK2: "deepseek2",
399400
MODEL_ARCH.CHATGLM: "chatglm",
400401
MODEL_ARCH.BITNET: "bitnet",
402+
MODEL_ARCH.BITNET_25: "bitnet-25",
401403
MODEL_ARCH.T5: "t5",
402404
MODEL_ARCH.T5ENCODER: "t5encoder",
403405
MODEL_ARCH.JAIS: "jais",
@@ -1162,6 +1164,28 @@ class MODEL_TENSOR(IntEnum):
11621164
MODEL_TENSOR.ATTN_SUB_NORM,
11631165
MODEL_TENSOR.FFN_SUB_NORM,
11641166
],
1167+
MODEL_ARCH.BITNET_25: [
1168+
MODEL_TENSOR.TOKEN_EMBD,
1169+
MODEL_TENSOR.OUTPUT_NORM,
1170+
MODEL_TENSOR.OUTPUT,
1171+
MODEL_TENSOR.ROPE_FREQS,
1172+
MODEL_TENSOR.ATTN_NORM,
1173+
MODEL_TENSOR.ATTN_Q,
1174+
MODEL_TENSOR.ATTN_K,
1175+
MODEL_TENSOR.ATTN_V,
1176+
MODEL_TENSOR.ATTN_OUT,
1177+
MODEL_TENSOR.ATTN_ROT_EMBD,
1178+
MODEL_TENSOR.FFN_GATE_INP,
1179+
MODEL_TENSOR.FFN_NORM,
1180+
MODEL_TENSOR.FFN_GATE,
1181+
MODEL_TENSOR.FFN_DOWN,
1182+
MODEL_TENSOR.FFN_UP,
1183+
MODEL_TENSOR.FFN_GATE_EXP,
1184+
MODEL_TENSOR.FFN_DOWN_EXP,
1185+
MODEL_TENSOR.FFN_UP_EXP,
1186+
MODEL_TENSOR.ATTN_SUB_NORM,
1187+
MODEL_TENSOR.FFN_SUB_NORM,
1188+
],
11651189
MODEL_ARCH.T5: [
11661190
MODEL_TENSOR.TOKEN_EMBD,
11671191
MODEL_TENSOR.OUTPUT,

gguf-py/gguf/tensor_mapping.py

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -141,6 +141,7 @@ class TensorNameMap:
141141
"model.layers.{bid}.self_attn.qkv_proj", # phi3
142142
"encoder.layers.{bid}.self_attention.query_key_value", # chatglm
143143
"transformer.layers.{bid}.attn.qkv_proj", # openelm
144+
"layers.{bid}.attention.wqkv",
144145
),
145146

146147
# Attention query
@@ -563,10 +564,14 @@ class TensorNameMap:
563564

564565
MODEL_TENSOR.ATTN_SUB_NORM: (
565566
"model.layers.{bid}.self_attn.inner_attn_ln", # bitnet
567+
"layers.{bid}.attention.attn_sub_norm", # bitnet
568+
"model.layers.{bid}.self_attn.attn_sub_norm",
566569
),
567570

568571
MODEL_TENSOR.FFN_SUB_NORM: (
569572
"model.layers.{bid}.mlp.ffn_layernorm", # bitnet
573+
"layers.{bid}.feed_forward.ffn_sub_norm", # bitnet
574+
"model.layers.{bid}.mlp.ffn_sub_norm",
570575
),
571576

572577
MODEL_TENSOR.DEC_ATTN_NORM: (

0 commit comments

Comments
 (0)