CUDA: 4D FlashAttention support#14628
Merged
ggerganov merged 2 commits intoggml-org:gg/llama-high-throughputfrom Jul 11, 2025
Merged
CUDA: 4D FlashAttention support#14628ggerganov merged 2 commits intoggml-org:gg/llama-high-throughputfrom
ggerganov merged 2 commits intoggml-org:gg/llama-high-throughputfrom