Successfully built, but encountering errors during building and generating gibberish #808
Unanswered
knguyen298
asked this question in
Q&A
Replies: 2 comments 4 replies
-
This is very interesting, but why you're using such an old software? Have you tried to upgrade? Below is my config and I am having about zero issues -- the software is crazy stable.
|
Beta Was this translation helpful? Give feedback.
2 replies
-
BTW what is your build command? Here is the example: make.sh
|
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
While I'm able to build ik_llama to the point where it runs, I'm not able to get meaningful generation with any model. During building, I'm encountering the following messages:
The following log output appears a few times while building
mmq-instance-iq4_ks_id.cu.o
:Despite these messages, I can still run the server and access the web GUI and API. Both methods of input return garbage text, using various different models (GPT-OSS 20B/120B and GLM 4.5 Air) and both special ik_llama quants and normal quants. Adjusting parameters does not fix the issue. The model files work fine in normal llama.cpp. llama.cpp is not installed system-wide. I have pulled the repo so all files are updated.
Specs:
Cuda compilation tools, release 12.0, V12.0.140 Build cuda_12.0.r12.0/compiler.32267302_0
gcc (Ubuntu 13.3.0-6ubuntu2~24.04) 13.3.0
-Nvidia driver version: 570.172.08
Any assistance would be appreciated - the garbage text that does generate does generate much faster than llama.cpp, so looking forward to using it in a functioning state!
Beta Was this translation helpful? Give feedback.
All reactions