llama.cpp supports ROCm, which would open this project to be used on AMD hardware as well: https://github.com/ggerganov/llama.cpp#hipblas