@@ -148,6 +148,7 @@ class TensorNameMap:
148148 "model.layers.{bid}.operator_norm" , # lfm2
149149 "model.transformer.blocks.{bid}.attn_norm" , # llada
150150 "layers.{bid}.input_layernorm" , # qwen3-embedding
151+ "model.layers.{bid}.attention_layernorm" # apertus
151152 ),
152153
153154 # Attention norm 2
@@ -325,6 +326,7 @@ class TensorNameMap:
325326 "model.layers.layers.{bid}.pre_mlp_norm" , # plamo2
326327 "model.transformer.blocks.{bid}.ff_norm" , # llada
327328 "layers.{bid}.post_attention_layernorm" , # qwen3-embedding
329+ "model.layers.{bid}.feedforward_layernorm" , # apertus
328330 ),
329331
330332 # Post feed-forward norm
@@ -535,6 +537,7 @@ class TensorNameMap:
535537 "transformer.layers.{bid}.attn.q_norm" , # openelm
536538 "model.layers.layers.{bid}.mixer.q" , # plamo2
537539 "layers.{bid}.self_attn.q_norm" , # qwen3-embedding
540+ "model.layers.{bid}.attention.query_layernorm" , # apertus
538541 ),
539542
540543 MODEL_TENSOR .ATTN_K_NORM : (
@@ -548,6 +551,7 @@ class TensorNameMap:
548551 "transformer.layers.{bid}.attn.k_norm" , # openelm
549552 "model.layers.layers.{bid}.mixer.k" , # plamo2
550553 "layers.{bid}.self_attn.k_norm" , # qwen3-embedding
554+ "model.layers.{bid}.attention.key_layernorm" , # apertus
551555 ),
552556
553557 MODEL_TENSOR .ROPE_FREQS : (
0 commit comments