Skip to content

Releases: JamePeng/llama-cpp-python

v0.3.17-cu130-AVX2-win-20251209

09 Dec 18:01

Choose a tag to compare

v0.3.17-cu130-AVX2-linux-20251209

09 Dec 17:32

Choose a tag to compare

v0.3.17-cu128-AVX2-win-20251209

09 Dec 18:16

Choose a tag to compare

v0.3.17-cu128-AVX2-linux-20251209

09 Dec 15:39

Choose a tag to compare

v0.3.17-cu126-AVX2-win-20251209

09 Dec 19:12

Choose a tag to compare

v0.3.17-cu126-AVX2-linux-20251209

09 Dec 15:11

Choose a tag to compare

v0.3.17-cu124-AVX2-win-20251209

09 Dec 19:03

Choose a tag to compare

v0.3.17-cu124-AVX2-linux-20251209

09 Dec 15:10

Choose a tag to compare

v0.3.17-cu130-Basic-win-20251207

07 Dec 19:28

Choose a tag to compare

feat: perf: optimize LlamaModel.metadata reading performance

  • Increase initial buffer size to 16KB to eliminate re-allocations for large chat templates.
  • Cache ctypes function references to reduce loop overhead.
  • Repeated model loading can result in a cumulative speed improvement of 1-3%.

feat: Update Submodule vendor/llama.cpp d9e03db..0a540f9
feat: Sync ggml-zendnn : add ZenDNN backend for AMD CPUs
feat: workflow: Added workflows for compiling with CUDA 13.0.2 on Windows and Linux.
feat: feat: Added the scan path for CUDA 13.0+ dynamic link libraries under Windows system ($env:CUDA_PATH\bin\x64)

v0.3.17-cu130-Basic-linux-20251207

07 Dec 16:54

Choose a tag to compare

feat: perf: optimize LlamaModel.metadata reading performance

  • Increase initial buffer size to 16KB to eliminate re-allocations for large chat templates.
  • Cache ctypes function references to reduce loop overhead.
  • Repeated model loading can result in a cumulative speed improvement of 1-3%.

feat: Update Submodule vendor/llama.cpp d9e03db..0a540f9
feat: Sync ggml-zendnn : add ZenDNN backend for AMD CPUs
feat: workflow: Added workflows for compiling with CUDA 13.0.2 on Windows and Linux.
feat: feat: Added the scan path for CUDA 13.0+ dynamic link libraries under Windows system ($env:CUDA_PATH\bin\x64)