Skip to content
Discussion options

You must be logged in to vote

Updating for whomever needs it: The llama-cli and other binaries in the root folder of the project are not symlinks to the recently built ones. So the llama-cli will never match with the one you just built after running the build commands.

Just build it yourself and run it from the build folder and it should work with the quantized models.

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by objectivecosta
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
1 participant