Skip to content
Discussion options

You must be logged in to vote

Meta-Llama-3-8B-F16.gguf looks like a base model, so it won't understand the prompt format that you are using. Try to use an instruction tuned variant like Meta-Llama-3-8B-Instruct-F16.gguf. Also, change -c to a power of 2: e.g. -c 2048

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@dgbaenar
Comment options

Answer selected by dgbaenar
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants