Skip to content

Commit 1be2787

Browse files
committed
Not calculating n_embd_head_k,v by n_embd / n_head
1 parent 07b55f4 commit 1be2787

File tree

1 file changed

+5
-1
lines changed

1 file changed

+5
-1
lines changed

src/llama-model.cpp

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1076,7 +1076,11 @@ void llama_model::load_hparams(llama_model_loader & ml) {
10761076
}
10771077
break;
10781078
default: type = LLM_TYPE_UNKNOWN;
1079-
}
1079+
}
1080+
1081+
// Load attention parameters
1082+
ml.get_key(LLM_KV_ATTENTION_KEY_LENGTH, hparams.n_embd_head_k, false);
1083+
ml.get_key(LLM_KV_ATTENTION_VALUE_LENGTH, hparams.n_embd_head_v, false);
10801084
} break;
10811085
case LLM_ARCH_GPT2:
10821086
{

0 commit comments

Comments
 (0)