Skip to content

Conversation

@jan-service-account
Copy link

Updates dev branch with latest release (b5598) from ggml-org/llama.cpp

jeffbolznv and others added 8 commits June 5, 2025 07:17
…ml-org#14006)

* memory : merge llama_kv_cache into llama_memory + new `llama_memory` API

ggml-ci

* context : fix casts

ggml-ci
Replace CMAKE_CUDA_ARCHITECTURES=native with nvidia-smi detection
as 'native' fails on autodl cloud environments.

Co-authored-by: pockers21 <[email protected]>
…ggml-org#14001)

* allowing B580 and U9-288V

* experimenting code to detect Xe2

* allowing coopmat only for Xe2 GPUs

* fixed comment wording

* fixed comment wording

* removed unnecessary driver check
@jan-service-account jan-service-account merged commit 55229f4 into dev Jun 6, 2025
9 checks passed
@jan-service-account jan-service-account deleted the update-dev-from-master-2025-06-06-00-08 branch June 6, 2025 00:19
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

9 participants