@@ -209,10 +209,6 @@ struct clip_layer {
209209 struct ggml_tensor * ff_gate_b = nullptr ;
210210 struct ggml_tensor * ff_down_w = nullptr ;
211211 struct ggml_tensor * ff_down_b = nullptr ;
212-
213- // post-ffn norm (output layer norm)
214- struct ggml_tensor * post_ffn_norm_w = nullptr ;
215- struct ggml_tensor * post_ffn_norm_b = nullptr ;
216212};
217213
218214struct clip_vision_model {
@@ -943,12 +939,6 @@ static ggml_cgraph * clip_image_build_graph_llama4(clip_ctx * ctx, const clip_im
943939 // residual 2
944940 cur = ggml_add (ctx0, embeddings, cur);
945941
946- // norm output
947- {
948- cur = ggml_norm (ctx0, cur, eps);
949- cur = ggml_add (ctx0, ggml_mul (ctx0, cur, model.layers [il].post_ffn_norm_w ), model.layers [il].post_ffn_norm_b );
950- }
951-
952942 embeddings = cur;
953943 }
954944
@@ -2041,7 +2031,7 @@ struct clip_model_loader {
20412031 vision_model.patch_embeddings_0 = get_tensor (TN_PATCH_EMBD, false );
20422032 vision_model.patch_embeddings_1 = get_tensor (TN_PATCH_EMBD_1, false );
20432033
2044- vision_model.position_embeddings = get_tensor (string_format ( TN_POS_EMBD, " v " ) , false );
2034+ vision_model.position_embeddings = get_tensor (TN_POS_EMBD, false );
20452035
20462036 // layers
20472037 vision_model.layers .resize (vision_model.hparams .n_layer );
@@ -2060,9 +2050,6 @@ struct clip_model_loader {
20602050 layer.ln_1_b = get_tensor (string_format (TN_LN_1, " v" , il, " bias" ), false );
20612051 layer.ln_2_b = get_tensor (string_format (TN_LN_2, " v" , il, " bias" ), false );
20622052
2063- layer.post_ffn_norm_b = get_tensor (string_format (TN_FFN_POST_NORM, " v" , il, " bias" ), false );
2064- layer.post_ffn_norm_w = get_tensor (string_format (TN_FFN_POST_NORM, " v" , il, " weight" ), false );
2065-
20662053 // new naming
20672054 layer.ff_up_w = get_tensor (string_format (TN_FFN_UP, " v" , il, " weight" ));
20682055 layer.ff_up_b = get_tensor (string_format (TN_FFN_UP, " v" , il, " bias" ), false );
0 commit comments