### Prerequisites - [X] I am running the latest code. Mention the version if possible as well. - [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.md). - [X] I searched using keywords relevant to my issue to make sure that I am creating a new issue that is not already open (or closed). - [X] I reviewed the [Discussions](https://github.com/ggerganov/llama.cpp/discussions), and have a new and useful enhancement to share. ### Feature Description Hi, seeing https://github.com/ggerganov/llama.cpp/releases provides a variety of windows prebuilt binary variants with CUDA, SYCL, vulkan etc.. like the latest vulkan enabled one: https://github.com/ggerganov/llama.cpp/releases/download/b4282/llama-b4282-bin-win-vulkan-x64.zip others like AVX512: [llama-b4282-bin-win-avx512-x64.zip](https://github.com/ggerganov/llama.cpp/releases/download/b4282/llama-b4282-bin-win-avx512-x64.zip) sadly no variant for Ubuntu is provided only: [llama-b4282-bin-ubuntu-x64.zip](https://github.com/ggerganov/llama.cpp/releases/download/b4282/llama-b4282-bin-ubuntu-x64.zip) which I downloaded and shows: ``` ./llama-cli --list-devices Available devices: ``` ### Motivation Vulkan seems for Linux right now is the most performant crossvendor variant specially with cooperative matrix support merged: https://github.com/ggerganov/llama.cpp/pull/10597 seems sad no easy testing and to have to build from source.. ideally could request more: CUDA, and HIP/RocM builds could be provided.. and even more than x64 buils also ubuntu arm64 builds.. ### Possible Implementation _No response_