Replies: 17 comments 10 replies
-
We're refactoring a few things to more easily support different architectures, though I haven't worked with AMD cards yet. Will try adding them to the onnx providers but may need some community testing to get right |
Beta Was this translation helpful? Give feedback.
-
Would ONNX work on the intel Arc GPUs? |
Beta Was this translation helpful? Give feedback.
-
I've got it to work with rocm. The problem is that the docker image becomes huge: Around 22G when I use https://github.com/bgs4free/Kokoro-FastAPI/tree/add-rocm-support |
Beta Was this translation helpful? Give feedback.
-
|
Beta Was this translation helpful? Give feedback.
-
What version of rocm did you get it working with I have a 5700xt and that thing only supports rocm 5.2 unofficially. Any chance you could get it working on that? Or maybe get direct-ml working with the docker file. |
Beta Was this translation helpful? Give feedback.
-
The I'm using the pytorch index for 6.2, because I had issues with the 6.3 version (version mismatch seems to be no issue AFAIK). Change it to 5.2, try and see if it works. Make sure you have enough disk space.
|
Beta Was this translation helpful? Give feedback.
-
Okay I tried that and it gave me the following error Acording to this thread https://stackoverflow.com/questions/73229163/amd-rocm-with-pytorch-on-navi10-rx-5700-rx-5700-xt you have to Set the environment variable.HSA_OVERRIDE_GFX_VERSION=10.3.0 I just added the variable to the docker file under env. Any idea what else i could try? |
Beta Was this translation helpful? Give feedback.
-
devices get passed through from the host. See: https://github.com/bgs4free/Kokoro-FastAPI/blob/add-rocm-support/docker/rocm/docker-compose.yml#L20 What is your host system? This was tested on Linux and I can't speak for any other OS. |
Beta Was this translation helpful? Give feedback.
-
Oh, okay its windows. I thought that wouldn't reaööy matter since the docker runs an Ubuntu image. |
Beta Was this translation helpful? Give feedback.
-
https://rocm.docs.amd.com/projects/install-on-linux/en/latest/how-to/docker.html Don't see anything about docker for windows here. |
Beta Was this translation helpful? Give feedback.
-
I checked my docker config it is running through wsl so that is technically Ubuntu so it should work. |
Beta Was this translation helpful? Give feedback.
-
Taking a shot at this shortly, will post links if anyone is able to test it out (I don't have an AMD card unfortunately, so can just skeleton it up) |
Beta Was this translation helpful? Give feedback.
-
My fork works on my linux machine. Didn't make a PR, though, because I personally didn't find it convincing enough to pursue. Please feel free to use my stuff if you see any value. |
Beta Was this translation helpful? Give feedback.
-
But for Arc I already implemented a version with ipex (intel extention for pytorch) support that uses nearly the exact code as CUDA GPUs For more on Arc please use Issue#106 |
Beta Was this translation helpful? Give feedback.
-
I'm interested in setting this up and would be happy to help test things out. I'm running Ubuntu 24.04.1 with an RX 7900 XTX. Thanks @bgs4free for your fork and rocm docker-compose file! Unfortunately I'm also seeing the issue where neither of the GPU devices The devices show up on the host outside the container, and I can at least get the container running by either removing the device flags or setting the privileged flag, but in both cases the devices aren't visible inside. Inside the container, I'm not entirely certain I have the right drivers installed, but ollama and alltalk_tts can both run and use my GPU. |
Beta Was this translation helpful? Give feedback.
-
I have built a Docker image containing the latest (Marc 23, 2025) version of master. This built against my fork of kokoro-fastapi, I will be maintaining this fork a bit. You can run the container via:
@bgs4free I have updated your base ROCm image to use the upstream rocm/pytorch images, which will get us around the 6.2 whl files. |
Beta Was this translation helpful? Give feedback.
-
I was able to make it work on my RX 7900xtx on ArchLinux using uv without docker
|
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hey, it'd be really nice if we could run this model API on AMD cards, and it would provide a very nice speedup over CPU. Do you think you could add this when you have some free time? It shouldn't be too hard since afaik ONNX models can run with ROCM just fine.
Beta Was this translation helpful? Give feedback.
All reactions