Skip to content

Releases: Yangxiaoz/llama.cpp

b5590

05 Jun 03:57
0d39844

Choose a tag to compare

ggml-vulkan: adds support for op CONV_TRANSPOSE_1D (#13813)

* * ggml-vulkan: adds op CONV_TRANSPOSE_1D

* test-backend-ops: adds more spohisticated tests for CONV_TRANSPOSE_1D

* Missing barrier added to shader.
Number of additional tests reduced to 108.

* * Fixes typo in variable name.

* Removes extra whitespaces.

* Adds int64->int32 casts to prevent possible warnings.

* Problem size reduced in tests to pass tests with llvmpipe.

* supports_op condition moved from unintended position

b5555

31 May 13:52
803f8ba

Choose a tag to compare

llama : deprecate explicit kv_self defrag/update calls (#13921)

ggml-ci

b5536

29 May 15:18
2b13162

Choose a tag to compare

gguf-py : add support for sub_type (in arrays) in GGUFWriter add_key_…

b5520

28 May 14:36

Choose a tag to compare

CUDA: add a flag "GGML_CUDA_JETSON_DEVICE" for optimization(#13856)

b5519

28 May 13:42
a682474

Choose a tag to compare

CUDA: fix FA tg at long context for CC >= 8.9 (#13852)

b5379

14 May 12:47
360a9c9

Choose a tag to compare

server : fix cache_tokens bug with no cache_prompt (#13533)