Skip to content

Commit 6c0715b

Browse files
committed
fix: update callback for ffn_moe_weighted and add callback for attn_out in deepseek2 model
1 parent 1e08157 commit 6c0715b

File tree

2 files changed

+2
-1
lines changed

2 files changed

+2
-1
lines changed

src/llama-graph.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1106,7 +1106,7 @@ ggml_tensor * llm_graph_context::build_moe_ffn(
11061106

11071107
if (!weight_before_ffn) {
11081108
experts = ggml_mul(ctx0, experts, weights);
1109-
cb(cur, "ffn_moe_weighted", il);
1109+
cb(experts, "ffn_moe_weighted", il);
11101110
}
11111111

11121112
ggml_tensor * cur_experts[LLAMA_MAX_EXPERTS] = { nullptr };

src/models/deepseek2.cpp

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -74,6 +74,7 @@ llm_build_deepseek2::llm_build_deepseek2(const llama_model & model, const llm_gr
7474
cur = build_attn(inp_attn,
7575
model.layers[il].wo, NULL,
7676
Qcur, Kcur, Vcur, nullptr, nullptr, nullptr, kq_scale, il);
77+
cb(cur, "attn_out", il);
7778
}
7879
else {
7980
ggml_tensor * q = NULL;

0 commit comments

Comments
 (0)