We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 3920faa commit 4d66bdcCopy full SHA for 4d66bdc
src/llama-model.cpp
@@ -1511,6 +1511,11 @@ void llama_model::load_hparams(llama_model_loader & ml) {
1511
ml.get_key(LLM_KV_EXPERT_FEED_FORWARD_LENGTH, hparams.n_ff_exp);
1512
ml.get_key(LLM_KV_EXPERT_SHARED_FEED_FORWARD_LENGTH, hparams.n_ff_shexp);
1513
1514
+ // TODO: read from gguf
1515
+ float n_dim = hparams.n_embd_head_k;
1516
+ float alpha = 1000.0f; // NTK-Aware
1517
+ hparams.rope_freq_base_train = 10000.0f * std::powf(alpha, n_dim / (n_dim - 2.0f));
1518
+
1519
switch (hparams.n_layer) {
1520
case 32: type = LLM_TYPE_A13B; break;
1521
default: type = LLM_TYPE_UNKNOWN;
0 commit comments