Skip to content

Commit 7f5d8a9

Browse files
committed
Fix loading weights
1 parent b164aa1 commit 7f5d8a9

File tree

1 file changed

+5
-0
lines changed

1 file changed

+5
-0
lines changed

src/llama-model.cpp

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1077,6 +1077,10 @@ void llama_model::load_hparams(llama_model_loader & ml) {
10771077
break;
10781078
default: type = LLM_TYPE_UNKNOWN;
10791079
}
1080+
1081+
// Load attention parameters
1082+
ml.get_key(LLM_KV_ATTENTION_KEY_LENGTH, hparams.n_embd_head_k);
1083+
ml.get_key(LLM_KV_ATTENTION_VALUE_LENGTH, hparams.n_embd_head_v);
10801084
} break;
10811085
case LLM_ARCH_GPT2:
10821086
{
@@ -17521,6 +17525,7 @@ struct llm_build_plamo2 : public llm_graph_context_mamba {
1752117525
const int64_t n_embd_head_q = hparams.n_embd_head_k;
1752217526
const int64_t n_embd_head_k = hparams.n_embd_head_k;
1752317527
const int64_t n_embd_head_v = hparams.n_embd_head_v;
17528+
int32_t n_head = hparams.n_head(il);
1752417529
int32_t n_head_kv = hparams.n_head_kv(il);
1752517530

1752617531
const int64_t q_offset = 0;

0 commit comments

Comments
 (0)