Skip to content

Commit 90f052d

Browse files
committed
Add tensor mappings for Apertus to global list instead
1 parent 5a02bd4 commit 90f052d

File tree

1 file changed

+4
-0
lines changed

1 file changed

+4
-0
lines changed

gguf-py/gguf/tensor_mapping.py

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -148,6 +148,7 @@ class TensorNameMap:
148148
"model.layers.{bid}.operator_norm", # lfm2
149149
"model.transformer.blocks.{bid}.attn_norm", # llada
150150
"layers.{bid}.input_layernorm", # qwen3-embedding
151+
"model.layers.{bid}.attention_layernorm" # apertus
151152
),
152153

153154
# Attention norm 2
@@ -325,6 +326,7 @@ class TensorNameMap:
325326
"model.layers.layers.{bid}.pre_mlp_norm", # plamo2
326327
"model.transformer.blocks.{bid}.ff_norm", # llada
327328
"layers.{bid}.post_attention_layernorm", # qwen3-embedding
329+
"model.layers.{bid}.feedforward_layernorm", # apertus
328330
),
329331

330332
# Post feed-forward norm
@@ -535,6 +537,7 @@ class TensorNameMap:
535537
"transformer.layers.{bid}.attn.q_norm", # openelm
536538
"model.layers.layers.{bid}.mixer.q", # plamo2
537539
"layers.{bid}.self_attn.q_norm", # qwen3-embedding
540+
"model.layers.{bid}.attention.query_layernorm", # apertus
538541
),
539542

540543
MODEL_TENSOR.ATTN_K_NORM: (
@@ -548,6 +551,7 @@ class TensorNameMap:
548551
"transformer.layers.{bid}.attn.k_norm", # openelm
549552
"model.layers.layers.{bid}.mixer.k", # plamo2
550553
"layers.{bid}.self_attn.k_norm", # qwen3-embedding
554+
"model.layers.{bid}.attention.key_layernorm", # apertus
551555
),
552556

553557
MODEL_TENSOR.ROPE_FREQS: (

0 commit comments

Comments
 (0)