Skip to content

Commit 6303ea2

Browse files
authored
initial jina-embeddings-v3 support
1 parent 8e186ef commit 6303ea2

File tree

2 files changed

+16
-0
lines changed

2 files changed

+16
-0
lines changed

gguf-py/gguf/constants.py

Lines changed: 14 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -272,6 +272,7 @@ class MODEL_ARCH(IntEnum):
272272
NOMIC_BERT = auto()
273273
NOMIC_BERT_MOE = auto()
274274
JINA_BERT_V2 = auto()
275+
JINA_BERT_V3 = auto()
275276
BLOOM = auto()
276277
STABLELM = auto()
277278
QWEN = auto()
@@ -534,6 +535,7 @@ class MODEL_TENSOR(IntEnum):
534535
MODEL_ARCH.NOMIC_BERT: "nomic-bert",
535536
MODEL_ARCH.NOMIC_BERT_MOE: "nomic-bert-moe",
536537
MODEL_ARCH.JINA_BERT_V2: "jina-bert-v2",
538+
MODEL_ARCH.JINA_BERT_V3: "jina-bert-v3",
537539
MODEL_ARCH.BLOOM: "bloom",
538540
MODEL_ARCH.STABLELM: "stablelm",
539541
MODEL_ARCH.QWEN: "qwen",
@@ -1020,6 +1022,18 @@ class MODEL_TENSOR(IntEnum):
10201022
MODEL_TENSOR.LAYER_OUT_NORM,
10211023
MODEL_TENSOR.CLS,
10221024
],
1025+
MODEL_ARCH.JINA_BERT_V3: [
1026+
MODEL_TENSOR.TOKEN_EMBD,
1027+
MODEL_TENSOR.TOKEN_EMBD_NORM,
1028+
MODEL_TENSOR.TOKEN_TYPES,
1029+
MODEL_TENSOR.OUTPUT_NORM,
1030+
MODEL_TENSOR.ATTN_OUT_NORM,
1031+
MODEL_TENSOR.ATTN_QKV,
1032+
MODEL_TENSOR.ATTN_OUT,
1033+
MODEL_TENSOR.FFN_DOWN,
1034+
MODEL_TENSOR.FFN_UP,
1035+
MODEL_TENSOR.LAYER_OUT_NORM,
1036+
],
10231037
MODEL_ARCH.MPT: [
10241038
MODEL_TENSOR.TOKEN_EMBD,
10251039
MODEL_TENSOR.OUTPUT_NORM,

gguf-py/gguf/tensor_mapping.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -157,6 +157,7 @@ class TensorNameMap:
157157
"h.{bid}.attn.c_attn", # gpt2
158158
"transformer.h.{bid}.mixer.Wqkv", # phi2
159159
"encoder.layers.{bid}.attn.Wqkv", # nomic-bert
160+
"encoder.layers.{bid}.mixer.Wqkv", # jina-bert-v3
160161
"model.layers.{bid}.self_attn.qkv_proj", # phi3
161162
"encoder.layers.{bid}.self_attention.query_key_value", # chatglm
162163
"transformer.layers.{bid}.attn.qkv_proj", # openelm
@@ -224,6 +225,7 @@ class TensorNameMap:
224225
"model.layers.layers.{bid}.self_attn.o_proj", # plamo
225226
"model.layers.{bid}.attention.wo", # internlm2
226227
"encoder.layers.{bid}.attn.out_proj", # nomic-bert
228+
"encoder.layers.{bid}.mixer.out_proj", # jina-bert-v3
227229
"transformer.decoder_layer.{bid}.multi_head_attention.linear", # Grok
228230
"transformer.blocks.{bid}.norm_attn_norm.attn.out_proj", # dbrx
229231
"encoder.layers.{bid}.self_attention.dense", # chatglm

0 commit comments

Comments
 (0)