From 5e67c4a576180cbc6c268182812df2fd8822892b Mon Sep 17 00:00:00 2001 From: Eric Curtin Date: Thu, 24 Oct 2024 15:38:41 +0100 Subject: [PATCH] Make Kompute error verbose about unsupported types Print the arch and quantization as strings. --- src/llama.cpp | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/llama.cpp b/src/llama.cpp index 24e1f1f01a857..4a6b0129bebc6 100644 --- a/src/llama.cpp +++ b/src/llama.cpp @@ -9127,7 +9127,7 @@ static int llama_model_load(const std::string & fname, llama_model & model, llam ) )) { // TODO(cebtenzzre): propagate this error outside of llama_load_model_from_file - LLAMA_LOG_WARN("%s: disabling Kompute due to unsupported model arch or quantization\n", __func__); + LLAMA_LOG_WARN("%s: disabling Kompute due to unsupported model arch (%s) or quantization (%s)\n", __func__, LLM_ARCH_NAMES.at(model.arch), llama_model_ftype_name(model.ftype).c_str()); params.n_gpu_layers = 0; } #endif