Skip to content

Commit 32a62d1

Browse files
committed
rm ffn_post_norm
1 parent 15605e4 commit 32a62d1

File tree

3 files changed

+2
-19
lines changed

3 files changed

+2
-19
lines changed

gguf-py/gguf/constants.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -492,7 +492,6 @@ class MODEL_TENSOR(IntEnum):
492492
V_ENC_FFN_UP = auto()
493493
V_ENC_FFN_GATE = auto()
494494
V_ENC_FFN_DOWN = auto()
495-
V_ENC_FFN_POST_NORM = auto()
496495
V_PRE_NORM = auto()
497496
V_POST_NORM = auto()
498497
V_MM_INP_NORM = auto()
@@ -751,7 +750,6 @@ class MODEL_TENSOR(IntEnum):
751750
MODEL_TENSOR.V_ENC_FFN_UP: "v.blk.{bid}.ffn_up",
752751
MODEL_TENSOR.V_ENC_FFN_GATE: "v.blk.{bid}.ffn_gate",
753752
MODEL_TENSOR.V_ENC_FFN_DOWN: "v.blk.{bid}.ffn_down",
754-
MODEL_TENSOR.V_ENC_FFN_POST_NORM: "v.blk.{bid}.ffn_post_norm",
755753
MODEL_TENSOR.V_PRE_NORM: "v.pre_ln",
756754
MODEL_TENSOR.V_POST_NORM: "v.post_ln",
757755
MODEL_TENSOR.V_MM_INP_PROJ: "mm.input_projection",
@@ -791,7 +789,6 @@ class MODEL_TENSOR(IntEnum):
791789
MODEL_TENSOR.V_ENC_FFN_UP,
792790
MODEL_TENSOR.V_ENC_FFN_GATE,
793791
MODEL_TENSOR.V_ENC_FFN_DOWN,
794-
MODEL_TENSOR.V_ENC_FFN_POST_NORM,
795792
MODEL_TENSOR.V_PRE_NORM,
796793
MODEL_TENSOR.V_POST_NORM,
797794
MODEL_TENSOR.V_MM_INP_PROJ,

tools/llava/clip-impl.h

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -47,7 +47,7 @@
4747
// tensor name constants
4848
//
4949

50-
#define TN_POS_EMBD "%s.position_embd.weight"
50+
#define TN_POS_EMBD "v.position_embd.weight"
5151
#define TN_CLASS_EMBD "v.class_embd"
5252
#define TN_PATCH_EMBD "v.patch_embd.weight" // not rename tensor with ".0" postfix for backwrad compat
5353
#define TN_PATCH_EMBD_1 "v.patch_embd.weight.1"
@@ -60,7 +60,6 @@
6060
#define TN_FFN_GATE "%s.blk.%d.ffn_gate.%s"
6161
#define TN_FFN_UP "%s.blk.%d.ffn_up.%s"
6262
#define TN_FFN_GATE "%s.blk.%d.ffn_gate.%s"
63-
#define TN_FFN_POST_NORM "%s.blk.%d.ffn_post_norm.%s"
6463
#define TN_LN_1 "%s.blk.%d.ln1.%s"
6564
#define TN_LN_2 "%s.blk.%d.ln2.%s"
6665
#define TN_LN_PRE "%s.pre_ln.%s"

tools/llava/clip.cpp

Lines changed: 1 addition & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -209,10 +209,6 @@ struct clip_layer {
209209
struct ggml_tensor * ff_gate_b = nullptr;
210210
struct ggml_tensor * ff_down_w = nullptr;
211211
struct ggml_tensor * ff_down_b = nullptr;
212-
213-
// post-ffn norm (output layer norm)
214-
struct ggml_tensor * post_ffn_norm_w = nullptr;
215-
struct ggml_tensor * post_ffn_norm_b = nullptr;
216212
};
217213

218214
struct clip_vision_model {
@@ -943,12 +939,6 @@ static ggml_cgraph * clip_image_build_graph_llama4(clip_ctx * ctx, const clip_im
943939
// residual 2
944940
cur = ggml_add(ctx0, embeddings, cur);
945941

946-
// norm output
947-
{
948-
cur = ggml_norm(ctx0, cur, eps);
949-
cur = ggml_add(ctx0, ggml_mul(ctx0, cur, model.layers[il].post_ffn_norm_w), model.layers[il].post_ffn_norm_b);
950-
}
951-
952942
embeddings = cur;
953943
}
954944

@@ -2041,7 +2031,7 @@ struct clip_model_loader {
20412031
vision_model.patch_embeddings_0 = get_tensor(TN_PATCH_EMBD, false);
20422032
vision_model.patch_embeddings_1 = get_tensor(TN_PATCH_EMBD_1, false);
20432033

2044-
vision_model.position_embeddings = get_tensor(string_format(TN_POS_EMBD, "v"), false);
2034+
vision_model.position_embeddings = get_tensor(TN_POS_EMBD, false);
20452035

20462036
// layers
20472037
vision_model.layers.resize(vision_model.hparams.n_layer);
@@ -2060,9 +2050,6 @@ struct clip_model_loader {
20602050
layer.ln_1_b = get_tensor(string_format(TN_LN_1, "v", il, "bias"), false);
20612051
layer.ln_2_b = get_tensor(string_format(TN_LN_2, "v", il, "bias"), false);
20622052

2063-
layer.post_ffn_norm_b = get_tensor(string_format(TN_FFN_POST_NORM, "v", il, "bias"), false);
2064-
layer.post_ffn_norm_w = get_tensor(string_format(TN_FFN_POST_NORM, "v", il, "weight"), false);
2065-
20662053
// new naming
20672054
layer.ff_up_w = get_tensor(string_format(TN_FFN_UP, "v", il, "weight"));
20682055
layer.ff_up_b = get_tensor(string_format(TN_FFN_UP, "v", il, "bias"), false);

0 commit comments

Comments
 (0)