File tree Expand file tree Collapse file tree 1 file changed +5
-0
lines changed Expand file tree Collapse file tree 1 file changed +5
-0
lines changed Original file line number Diff line number Diff line change @@ -1077,6 +1077,10 @@ void llama_model::load_hparams(llama_model_loader & ml) {
10771077 break;
10781078 default: type = LLM_TYPE_UNKNOWN;
10791079 }
1080+
1081+ // Load attention parameters
1082+ ml.get_key(LLM_KV_ATTENTION_KEY_LENGTH, hparams.n_embd_head_k);
1083+ ml.get_key(LLM_KV_ATTENTION_VALUE_LENGTH, hparams.n_embd_head_v);
10801084 } break;
10811085 case LLM_ARCH_GPT2:
10821086 {
@@ -17521,6 +17525,7 @@ struct llm_build_plamo2 : public llm_graph_context_mamba {
1752117525 const int64_t n_embd_head_q = hparams.n_embd_head_k;
1752217526 const int64_t n_embd_head_k = hparams.n_embd_head_k;
1752317527 const int64_t n_embd_head_v = hparams.n_embd_head_v;
17528+ int32_t n_head = hparams.n_head(il);
1752417529 int32_t n_head_kv = hparams.n_head_kv(il);
1752517530
1752617531 const int64_t q_offset = 0;
You can’t perform that action at this time.
0 commit comments