Skip to content

Conversation

ggerganov
Copy link
Member

@ngxson
Copy link
Collaborator

ngxson commented Oct 15, 2025

Does this fix it? #16595

Yes thanks, it works now

@ggerganov ggerganov marked this pull request as ready for review October 15, 2025 13:51
@ggerganov ggerganov requested a review from ngxson as a code owner October 15, 2025 13:51
@ggerganov ggerganov force-pushed the gg/server-fix-img-logs branch from 2f26d3a to d552d79 Compare October 15, 2025 13:53
@ggerganov ggerganov merged commit 17304cb into master Oct 15, 2025
2 checks passed
@ggerganov ggerganov deleted the gg/server-fix-img-logs branch October 15, 2025 13:53
yael-works pushed a commit to yael-works/llama.cpp that referenced this pull request Oct 15, 2025
gabe-l-hart added a commit to gabe-l-hart/llama.cpp that referenced this pull request Oct 15, 2025
* origin/master:
Add server-driven parameter defaults and syncing (ggml-org#16515)
metal: optimise `GGML_OP_SUM` (ggml-org#16559)
server : fix img token logs (ggml-org#16595)
llama-quant: add support for mmproj (ggml-org#16592)
CUDA: Changing the CUDA scheduling strategy to spin (ggml-org#16585)
server : fix mtmd checkpoints (ggml-org#16591)
metal : avoid using Metal's gpuAddress property (ggml-org#16576)
vulkan: Add ACC_TYPE_VEC2 implementation (ggml-org#16203)
CUDA + openCL: fix bug in accessing rms_norm->src while doing fusion (ggml-org#16577)
vulkan: Support FA with K/V in F32 (ggml-org#16543)
vulkan: Improve build time for MSVC (ggml-org#16545)
CUDA: enable FA for FP32 KV cache (ggml-org#16546)
CUDA: use fastdiv + ggml_cuda_mad for mmvf (ggml-org#16557)
CUDA: add fp kernel for larger batch size MoE (ggml-org#16512)
cuda : remove legacy copy-op pointer indirection code (ggml-org#16485)
server : dynamic token limit for prompt cache (ggml-org#16560)
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants