Releases: mofosyne/llama.cpp
Releases · mofosyne/llama.cpp
b2898
doc: add references to hugging face GGUF-my-repo quantisation web too…
b2886
script : sync ggml-rpc
b2876
llama : disable pipeline parallelism with nkvo (#7265)
b2874
Add left recursion check: quit early instead of going into an infinit…
b2866
convert.py: Outfile default name change and additional metadata suppo…
b2864
[SYCL] Add oneapi runtime dll files to win release package (#7241) * add oneapi running time dlls to release package * fix path * fix path * fix path * fix path * fix path --------- Co-authored-by: Zhang <[email protected]>
b2839
Fix memory bug in grammar parser (#7194)
The llama.cpp grammar parser had a bug where forgetting to add a closing
quotation mark to strings would cause parsing to crash. Anyone running a
server on a public endpoint is advised to upgrade. To reproduce this bug
    ./llamafile -m foo.gguf -p bar --grammar 'root::="'
Credit for discovering and reporting this issue goes to Eclypsium
Security Researcher Richard Johnson <[email protected]>.
b2836
Minor arithmetic improvement to mmvq wrapper kernel (#7172)
b2806
Further tidy on Android instructions README.md (#7077) * Further tidy on Android instructions README.md Fixed some logic when following readme direction * Clean up redundent information A new user arriving will see simple directions on llama.cpp homepage * corrected puncuation Period after cmake, colon after termux * re-word for clarity method seems to be more correct, instead of alternative in this context * Organized required packages per build type building llama.cpp with NDK on a pc doesn't require installing clang, cmake, git, or wget in termux. * README.md corrected title * fix trailing whitespace
b2794
Adding support for the --numa argument for llama-bench. (#7080)