Skip to content

Releases: ngxson/llama.cpp

b4967

27 Mar 02:38
f17a3bb
Compare
Choose a tag to compare
SYCL: implement memset ggml backend buffer interface (#12580)

* SYCL: implement memset ggml backend buffer interface

* use GGML_ABORT macro

* Do not wait for all queues to finish for memset operation

b4966

26 Mar 23:39
bd40678
Compare
Choose a tag to compare
HIP: Add support for RDNA4 targets (#12372)

b4964

26 Mar 18:49
2447ad8
Compare
Choose a tag to compare
upgrade to llguidance 0.7.10 (#12576)

b4963

26 Mar 14:49
02082f1
Compare
Choose a tag to compare
clip: Fix llama-llava-clip-quantize-cli quantization error under CUDA…

b4961

26 Mar 11:46
5ed38b6
Compare
Choose a tag to compare
ggml : fix MUL_MAT_ID repack with Q8_K (#12544)

* ggml : fix MUL_MAT_ID repack with Q8_K

ggml-ci

* ggml : improve repack templates

ggml-ci

b4958

25 Mar 18:47
ef19c71
Compare
Choose a tag to compare
run: de-duplicate fmt and format functions and optimize (#11596)

b4957

25 Mar 12:00
053b3f9
Compare
Choose a tag to compare
ggml-cpu : update KleidiAI to v1.5.0 (#12568)

ggml-cpu : bug fix related to KleidiAI LHS packing

Signed-off-by: Dan Johansson <[email protected]>

b4956

25 Mar 11:29
e2f5601
Compare
Choose a tag to compare
SYCL: disable Q4_0 reorder optimization (#12560)

ggml-ci

b4953

25 Mar 08:05
2d77d88
Compare
Choose a tag to compare
context : fix worst-case reserve outputs (#12545)

ggml-ci

b4951

24 Mar 17:04
2b65ae3
Compare
Choose a tag to compare
opencl: simplify kernel embedding logic in cmakefile (#12503)

Co-authored-by: Max Krasnyansky <[email protected]>