Skip to content

Commit b7675ea

Browse files
authored
fix embedding scale being applied twice
1 parent 94bcbbf commit b7675ea

File tree

1 file changed

+0
-2
lines changed

1 file changed

+0
-2
lines changed

src/llama-model.cpp

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6912,8 +6912,6 @@ struct llm_build_grok : public llm_graph_context {
69126912

69136913
inpL = build_inp_embd(model.tok_embd);
69146914

6915-
inpL = ggml_scale(ctx0, inpL, hparams.f_embedding_scale);
6916-
69176915
// inp_pos - contains the positions
69186916
ggml_tensor * inp_pos = build_inp_pos();
69196917

0 commit comments

Comments
 (0)