@@ -4072,23 +4072,21 @@ static void llm_load_hparams(
40724072 case LLM_ARCH_DEEPSEEK2:
40734073 {
40744074 if (hparams.n_head_kv() == 1) {
4075- printf("==========================================================================\n");
4076- printf("Detected incompatible DeepSeek model.\n");
4077- printf("Will try to fix, but there are no guarantees\n\n");
4078- printf("*** Your prompt processing speed will be crippled ***\n\n");
4079- printf("Consider making your own ik_llama.cpp compatible model or\n");
4080- printf("ask the model provider to make one for you,\n");
40814075 int n_nead_kv = hparams.n_gqa();
40824076 if (n_nead_kv%16 != 0 || hparams.n_embd_head_k != 576 || hparams.n_embd_head_v != 512 ||
40834077 hparams.n_rot != 64) {
4078+ printf("==========================================================================\n");
4079+ printf("Detected incompatible DeepSeek model without a known way to fixc it.\n");
4080+ printf("Consider making your own ik_llama.cpp compatible model or\n");
4081+ printf("ask the model provider to make one for you,\n\n");
40844082 printf("Sorry, uknown model => cannot fix it => bailing out\n");
4083+ printf("==========================================================================\n");
40854084 GGML_ABORT("Fatal error");
40864085 }
4086+ printf("================= Adjusted mainline llama.cpp MLA tensors to ik_llama.cpp\n");
40874087 for (auto& item : hparams.n_head_kv_arr) item = n_nead_kv;
40884088 hparams.n_embd_head_k = 192;
40894089 hparams.n_embd_head_v = 128;
4090- printf("==========================================================================\n");
4091- //GGML_ABORT("Fatal error");
40924090 }
40934091 bool is_lite = (hparams.n_layer == 27);
40944092 ml.get_key(LLM_KV_ATTENTION_LAYERNORM_RMS_EPS, hparams.f_norm_rms_eps);
0 commit comments