-
Notifications
You must be signed in to change notification settings - Fork 13.3k
Closed
Labels
Apple Metalhttps://en.wikipedia.org/wiki/Metal_(API)https://en.wikipedia.org/wiki/Metal_(API)bug-unconfirmed
Description
Name and Version
build/bin/llama-server --version
version: 6275 (4d917cd4)
built with Apple clang version 16.0.0 (clang-1600.0.26.6) for arm64-apple-darwin24.6.0
Operating systems
Mac
GGML backends
Metal
Hardware
Metal
Models
build/bin/llama-server -hf bartowski/mistralai_Mistral-Small-3.2-24B-Instruct-2506-GGUF:Q8_0 -v
Problem description & steps to reproduce
Hello, everyone,
Llama-server crashes with Mistrall-Small when I pass it an image for processing. Please, help me debug this further.
build/bin/llama-server -hf bartowski/mistralai_Mistral-Small-3.2-24B-Instruct-2506-GGUF:Q8_0 -v
...
Then pass prompt with an image.
First Bad Commit
No response
Relevant log output
process_chun: processing image...
encoding image slice...
/Users/bonk/proj/llama.cpp/ggml/src/ggml-metal/ggml-metal.m:4734: GGML_ASSERT(ggml_is_contiguous(src0)) failed
(lldb) process attach --pid 25022
Process 25022 stopped
* thread #1, queue = 'ggml-metal', stop reason = signal SIGSTOP
frame #0: 0x000000018f691204 libsystem_kernel.dylib`__wait4 + 8
libsystem_kernel.dylib`__wait4:
-> 0x18f691204 <+8>: b.lo 0x18f691224 ; <+40>
0x18f691208 <+12>: pacibsp
0x18f69120c <+16>: stp x29, x30, [sp, #-0x10]!
0x18f691210 <+20>: mov x29, sp
Target 0: (llama-server) stopped.
Executable module set to "/Users/dfadeyev/proj/llama.cpp/build/bin/llama-server".
Architecture set to: arm64-apple-macosx-.
(lldb) bt
* thread #1, queue = 'ggml-metal', stop reason = signal SIGSTOP
* frame #0: 0x000000018f691204 libsystem_kernel.dylib`__wait4 + 8
frame #1: 0x0000000102948614 libggml-base.dylib`ggml_abort + 156
frame #2: 0x0000000102c6c1c4 libggml-metal.dylib`ggml_metal_encode_node + 33088
frame #3: 0x0000000102c63fe8 libggml-metal.dylib`__ggml_backend_metal_set_n_cb_block_invoke + 588
frame #4: 0x000000018f52d874 libdispatch.dylib`_dispatch_client_callout2 + 16
frame #5: 0x000000018f5287b8 libdispatch.dylib`_dispatch_apply_serial + 164
frame #6: 0x000000018f52d85c libdispatch.dylib`_dispatch_client_callout + 16
frame #7: 0x000000018f523da8 libdispatch.dylib`_dispatch_sync_invoke_and_complete + 56
frame #8: 0x000000018f527e1c libdispatch.dylib`_dispatch_apply_with_attr_f + 1424
frame #9: 0x000000018f527f34 libdispatch.dylib`dispatch_apply + 96
frame #10: 0x0000000102c63ac8 libggml-metal.dylib`ggml_backend_metal_graph_compute + 444
frame #11: 0x0000000102963200 libggml-base.dylib`ggml_backend_sched_graph_compute_async + 2928
frame #12: 0x000000010296261c libggml-base.dylib`ggml_backend_sched_graph_compute + 24
frame #13: 0x000000010270c334 libmtmd.dylib`clip_image_batch_encode(clip_ctx*, int, clip_image_f32_batch const*, float*) + 5816
frame #14: 0x00000001026f7020 libmtmd.dylib`mtmd_encode + 344
frame #15: 0x000000010271d848 libmtmd.dylib`mtmd_helper_eval_chunk_single + 180
frame #16: 0x0000000102230d70 llama-server`server_tokens::process_chunk(llama_context*, mtmd_context*, int, int, int&) + 212
frame #17: 0x000000010222e9bc llama-server`server_context::update_slots() + 5324
frame #18: 0x00000001021aa178 llama-server`server_queue::start_loop() + 844
frame #19: 0x00000001021a74b8 llama-server`main + 20212
frame #20: 0x000000018f32ab98 dyld`start + 6076
(lldb) quit
Abort trap: 6
Metadata
Metadata
Assignees
Labels
Apple Metalhttps://en.wikipedia.org/wiki/Metal_(API)https://en.wikipedia.org/wiki/Metal_(API)bug-unconfirmed