Releases: Yangxiaoz/llama.cpp
Releases · Yangxiaoz/llama.cpp
b5590
ggml-vulkan: adds support for op CONV_TRANSPOSE_1D (#13813) * * ggml-vulkan: adds op CONV_TRANSPOSE_1D * test-backend-ops: adds more spohisticated tests for CONV_TRANSPOSE_1D * Missing barrier added to shader. Number of additional tests reduced to 108. * * Fixes typo in variable name. * Removes extra whitespaces. * Adds int64->int32 casts to prevent possible warnings. * Problem size reduced in tests to pass tests with llvmpipe. * supports_op condition moved from unintended position
b5555
llama : deprecate explicit kv_self defrag/update calls (#13921) ggml-ci
b5536
gguf-py : add support for sub_type (in arrays) in GGUFWriter add_key_…
b5520
CUDA: add a flag "GGML_CUDA_JETSON_DEVICE" for optimization(#13856)
b5519
CUDA: fix FA tg at long context for CC >= 8.9 (#13852)
b5379
server : fix cache_tokens bug with no cache_prompt (#13533)