forked from ggml-org/llama.cpp
-
Notifications
You must be signed in to change notification settings - Fork 580
Closed
Description
I would like to compile koboldcpp with rocWMMA support on gfx12.
I know that rocWMMA is not compatible with llamacpp and koboldcpp as of ROCm 6.4.1 but by installing rocWMMA from GitHub I can successful compile llamacpp with it and enjoy using flash attention and kv quantization.
I would like to know if it is possible to overwrite the standard behaviour and make koboldcpp compile with rocWMMA on gfx12.
Thanks in advance for the help!
OS:Linux (official Almalinux container for AMD HIP/ROCm)
GPU: Rx 9070 XT
Metadata
Metadata
Assignees
Labels
No labels