Skip to content

Commit 11c17d7

Browse files
committed
- tidying up code
- simplifying graph-context expecting both dense weights
1 parent e5c8808 commit 11c17d7

File tree

2 files changed

+9
-12
lines changed

2 files changed

+9
-12
lines changed

src/llama-graph.cpp

Lines changed: 7 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -1856,18 +1856,15 @@ llm_graph_input_mem_hybrid * llm_graph_context::build_inp_mem_hybrid() const {
18561856
void llm_graph_context::build_dense_out(
18571857
ggml_tensor * dense_2,
18581858
ggml_tensor * dense_3) const {
1859-
if (!cparams.embeddings) {
1859+
if (!cparams.embeddings || dense_2 == nullptr || dense_3 == nullptr) {
18601860
return;
18611861
}
1862-
ggml_tensor * cur = res->get_embd_pooled();
1863-
if (dense_2 != nullptr) {
1864-
cur = ggml_mul_mat(ctx0, dense_2, cur);
1865-
cb(cur, "result_embd_pooled", -1);
1866-
}
1867-
if (dense_3 != nullptr) {
1868-
cur = ggml_mul_mat(ctx0, dense_3, cur);
1869-
cb(cur, "result_embd_pooled", -1);
1870-
}
1862+
ggml_tensor * cur = res->t_embd_pooled != nullptr ? res->t_embd_pooled : res->t_embd;
1863+
GGML_ASSERT(cur != nullptr && "missing t_embd_pooled/t_embd");
1864+
1865+
cur = ggml_mul_mat(ctx0, dense_2, cur);
1866+
cur = ggml_mul_mat(ctx0, dense_3, cur);
1867+
cb(cur, "result_embd_pooled", -1);
18711868
res->t_embd_pooled = cur;
18721869
ggml_build_forward_expand(gf, cur);
18731870
}

src/llama-model.h

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -440,8 +440,8 @@ struct llama_model {
440440
//Dense linear projections for SentenceTransformers models like embeddinggemma
441441
// For Sentence Transformers models structure see
442442
// https://sbert.net/docs/sentence_transformer/usage/custom_models.html#structure-of-sentence-transformer-models
443-
ggml_tensor * dense_2_out_layers;
444-
ggml_tensor * dense_3_out_layers;
443+
struct ggml_tensor * dense_2_out_layers = nullptr;
444+
struct ggml_tensor * dense_3_out_layers = nullptr;
445445

446446
llama_model_params params;
447447

0 commit comments

Comments
 (0)