Skip to content

Commit 7bf48e3

Browse files
authored
Merge branch 'main' into fix/server_llama_call_thread_starvation
2 parents 9ec5460 + 7ecdd94 commit 7bf48e3

File tree

4 files changed

+26
-3
lines changed

4 files changed

+26
-3
lines changed

CHANGELOG.md

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,10 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
77

88
## [Unreleased]
99

10+
## [0.3.2]
11+
12+
- feat: Update llama.cpp to ggerganov/llama.cpp@74d73dc85cc2057446bf63cc37ff649ae7cebd80
13+
1014
## [0.3.1]
1115

1216
- feat: Update llama.cpp to ggerganov/llama.cpp@c919d5db39c8a7fcb64737f008e4b105ee0acd20

CMakeLists.txt

Lines changed: 20 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,6 +6,10 @@ option(LLAMA_BUILD "Build llama.cpp shared library and install alongside python
66
option(LLAVA_BUILD "Build llava shared library and install alongside python package" ON)
77

88
function(llama_cpp_python_install_target target)
9+
if(NOT TARGET ${target})
10+
return()
11+
endif()
12+
913
install(
1014
TARGETS ${target}
1115
LIBRARY DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/llama_cpp/lib
@@ -75,7 +79,22 @@ if (LLAMA_BUILD)
7579
add_subdirectory(vendor/llama.cpp)
7680
llama_cpp_python_install_target(llama)
7781
llama_cpp_python_install_target(ggml)
78-
82+
83+
llama_cpp_python_install_target(ggml-base)
84+
85+
llama_cpp_python_install_target(ggml-amx)
86+
llama_cpp_python_install_target(ggml-blas)
87+
llama_cpp_python_install_target(ggml-can)
88+
llama_cpp_python_install_target(ggml-cpu)
89+
llama_cpp_python_install_target(ggml-cuda)
90+
llama_cpp_python_install_target(ggml-hip)
91+
llama_cpp_python_install_target(ggml-kompute)
92+
llama_cpp_python_install_target(ggml-metal)
93+
llama_cpp_python_install_target(ggml-musa)
94+
llama_cpp_python_install_target(ggml-rpc)
95+
llama_cpp_python_install_target(ggml-sycl)
96+
llama_cpp_python_install_target(ggml-vulkan)
97+
7998
# Workaround for Windows + CUDA https://github.com/abetlen/llama-cpp-python/issues/563
8099
if (WIN32)
81100
install(

llama_cpp/__init__.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
11
from .llama_cpp import *
22
from .llama import *
33

4-
__version__ = "0.3.1"
4+
__version__ = "0.3.2"

vendor/llama.cpp

0 commit comments

Comments
 (0)