Releases: EmbeddedLLM/vllm-rocm
ROCm Wheels - 2025-11-21
Large ML framework wheels for vLLM ROCm
Included Wheels
cupy_cuda12x-13.6.0-cp312-cp312-manylinux2014_x86_64.whl
nvidia_cublas_cu12-12.8.4.1-py3-none-manylinux_2_27_x86_64.whl
nvidia_cudnn_cu12-9.10.2.21-py3-none-manylinux_2_27_x86_64.whl
nvidia_cufft_cu12-11.3.3.83-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_cusolver_cu12-11.7.3.90-py3-none-manylinux_2_27_x86_64.whl
nvidia_cusparse_cu12-12.5.8.93-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_cusparselt_cu12-0.7.1-py3-none-manylinux2014_x86_64.whl
nvidia_nccl_cu12-2.27.5-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_nvshmem_cu12-3.3.20-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
torch-2.9.0a0+git1c57644-cp312-cp312-linux_x86_64.whl
torch-2.9.1-cp312-cp312-manylinux_2_28_x86_64.whl
triton-3.4.0-cp312-cp312-linux_x86_64.whl
triton-3.5.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl
Build Info
- ROCm Version: 7.1
- Python Version: 3.12
- GPU Architectures: gfx942
- Commit: 035f0f1
Installation
These wheels are automatically referenced by the PyPI index at:
https://EmbeddedLLM.github.io/vllm-rocm/simple/
Use: pip install vllm-rocm --index-url https://EmbeddedLLM.github.io/vllm-rocm/simple/
ROCm Wheels - 2025-11-21
Large ML framework wheels for vLLM ROCm
Included Wheels
cupy_cuda12x-13.6.0-cp312-cp312-manylinux2014_x86_64.whl
nvidia_cublas_cu12-12.8.4.1-py3-none-manylinux_2_27_x86_64.whl
nvidia_cudnn_cu12-9.10.2.21-py3-none-manylinux_2_27_x86_64.whl
nvidia_cufft_cu12-11.3.3.83-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_cusolver_cu12-11.7.3.90-py3-none-manylinux_2_27_x86_64.whl
nvidia_cusparse_cu12-12.5.8.93-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_cusparselt_cu12-0.7.1-py3-none-manylinux2014_x86_64.whl
nvidia_nccl_cu12-2.27.5-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_nvshmem_cu12-3.3.20-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
torch-2.9.0a0+git1c57644-cp312-cp312-linux_x86_64.whl
torch-2.9.1-cp312-cp312-manylinux_2_28_x86_64.whl
triton-3.4.0-cp312-cp312-linux_x86_64.whl
triton-3.5.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl
Build Info
- ROCm Version: 7.1
- Python Version: 3.12
- GPU Architectures: gfx942
- Commit: 7318db8
Installation
These wheels are automatically referenced by the PyPI index at:
https://EmbeddedLLM.github.io/vllm-rocm/simple/
Use: pip install vllm-rocm --index-url https://EmbeddedLLM.github.io/vllm-rocm/simple/
ROCm Wheels - 2025-11-21
Large ML framework wheels for vLLM ROCm
Included Wheels
cupy_cuda12x-13.6.0-cp312-cp312-manylinux2014_x86_64.whl
nvidia_cublas_cu12-12.8.4.1-py3-none-manylinux_2_27_x86_64.whl
nvidia_cudnn_cu12-9.10.2.21-py3-none-manylinux_2_27_x86_64.whl
nvidia_cufft_cu12-11.3.3.83-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_cusolver_cu12-11.7.3.90-py3-none-manylinux_2_27_x86_64.whl
nvidia_cusparse_cu12-12.5.8.93-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_cusparselt_cu12-0.7.1-py3-none-manylinux2014_x86_64.whl
nvidia_nccl_cu12-2.27.5-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_nvshmem_cu12-3.3.20-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
torch-2.9.0a0+git1c57644-cp312-cp312-linux_x86_64.whl
torch-2.9.1-cp312-cp312-manylinux_2_28_x86_64.whl
triton-3.4.0-cp312-cp312-linux_x86_64.whl
triton-3.5.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl
Build Info
- ROCm Version: 7.1
- Python Version: 3.12
- GPU Architectures: gfx942
- Commit: a98d10f
Installation
These wheels are automatically referenced by the PyPI index at:
https://EmbeddedLLM.github.io/vllm-rocm/simple/
Use: pip install vllm-rocm --index-url https://EmbeddedLLM.github.io/vllm-rocm/simple/
ROCm Wheels - 2025-11-21
Large ML framework wheels for vLLM ROCm
Included Wheels
cupy_cuda12x-13.6.0-cp312-cp312-manylinux2014_x86_64.whl
nvidia_cublas_cu12-12.8.4.1-py3-none-manylinux_2_27_x86_64.whl
nvidia_cudnn_cu12-9.10.2.21-py3-none-manylinux_2_27_x86_64.whl
nvidia_cufft_cu12-11.3.3.83-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_cusolver_cu12-11.7.3.90-py3-none-manylinux_2_27_x86_64.whl
nvidia_cusparse_cu12-12.5.8.93-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_cusparselt_cu12-0.7.1-py3-none-manylinux2014_x86_64.whl
nvidia_nccl_cu12-2.27.5-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_nvshmem_cu12-3.3.20-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
torch-2.9.0a0+git1c57644-cp312-cp312-linux_x86_64.whl
torch-2.9.1-cp312-cp312-manylinux_2_28_x86_64.whl
triton-3.4.0-cp312-cp312-linux_x86_64.whl
triton-3.5.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl
Build Info
- ROCm Version: 7.1
- Python Version: 3.12
- GPU Architectures: gfx942
- Commit: 0a8bd9f
Installation
These wheels are automatically referenced by the PyPI index at:
https://EmbeddedLLM.github.io/vllm-rocm/simple/
Use: pip install vllm-rocm --index-url https://EmbeddedLLM.github.io/vllm-rocm/simple/
ROCm Wheels - 2025-11-21
Large ML framework wheels for vLLM ROCm
Included Wheels
cupy_cuda12x-13.6.0-cp312-cp312-manylinux2014_x86_64.whl
nvidia_cublas_cu12-12.8.4.1-py3-none-manylinux_2_27_x86_64.whl
nvidia_cudnn_cu12-9.10.2.21-py3-none-manylinux_2_27_x86_64.whl
nvidia_cufft_cu12-11.3.3.83-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_cusolver_cu12-11.7.3.90-py3-none-manylinux_2_27_x86_64.whl
nvidia_cusparse_cu12-12.5.8.93-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_cusparselt_cu12-0.7.1-py3-none-manylinux2014_x86_64.whl
nvidia_nccl_cu12-2.27.5-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_nvshmem_cu12-3.3.20-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
torch-2.9.0a0+git1c57644-cp312-cp312-linux_x86_64.whl
torch-2.9.1-cp312-cp312-manylinux_2_28_x86_64.whl
triton-3.4.0-cp312-cp312-linux_x86_64.whl
triton-3.5.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl
Build Info
- ROCm Version: 7.1
- Python Version: 3.12
- GPU Architectures: gfx942
- Commit: dea8943
Installation
These wheels are automatically referenced by the PyPI index at:
https://EmbeddedLLM.github.io/vllm-rocm/simple/
Use: pip install vllm-rocm --index-url https://EmbeddedLLM.github.io/vllm-rocm/simple/
ROCm Wheels - 2025-11-21
Large ML framework wheels for vLLM ROCm
Included Wheels
cupy_cuda12x-13.6.0-cp312-cp312-manylinux2014_x86_64.whl
nvidia_cublas_cu12-12.8.4.1-py3-none-manylinux_2_27_x86_64.whl
nvidia_cudnn_cu12-9.10.2.21-py3-none-manylinux_2_27_x86_64.whl
nvidia_cufft_cu12-11.3.3.83-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_cusolver_cu12-11.7.3.90-py3-none-manylinux_2_27_x86_64.whl
nvidia_cusparse_cu12-12.5.8.93-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_cusparselt_cu12-0.7.1-py3-none-manylinux2014_x86_64.whl
nvidia_nccl_cu12-2.27.5-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_nvshmem_cu12-3.3.20-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
torch-2.9.0a0+git1c57644-cp312-cp312-linux_x86_64.whl
torch-2.9.1-cp312-cp312-manylinux_2_28_x86_64.whl
triton-3.4.0-cp312-cp312-linux_x86_64.whl
triton-3.5.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl
Build Info
- ROCm Version: 7.1
- Python Version: 3.12
- GPU Architectures: gfx942
- Commit: 035f0f1
Installation
These wheels are automatically referenced by the PyPI index at:
https://EmbeddedLLM.github.io/vllm-rocm/simple/
Use: pip install vllm-rocm --index-url https://EmbeddedLLM.github.io/vllm-rocm/simple/
ROCm Wheels - 2025-11-21
Large ML framework wheels for vLLM ROCm
Included Wheels
cupy_cuda12x-13.6.0-cp312-cp312-manylinux2014_x86_64.whl
nvidia_cublas_cu12-12.8.4.1-py3-none-manylinux_2_27_x86_64.whl
nvidia_cudnn_cu12-9.10.2.21-py3-none-manylinux_2_27_x86_64.whl
nvidia_cufft_cu12-11.3.3.83-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_cusolver_cu12-11.7.3.90-py3-none-manylinux_2_27_x86_64.whl
nvidia_cusparse_cu12-12.5.8.93-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_cusparselt_cu12-0.7.1-py3-none-manylinux2014_x86_64.whl
nvidia_nccl_cu12-2.27.5-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
nvidia_nvshmem_cu12-3.3.20-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
torch-2.9.0a0+git1c57644-cp312-cp312-linux_x86_64.whl
torch-2.9.1-cp312-cp312-manylinux_2_28_x86_64.whl
triton-3.4.0-cp312-cp312-linux_x86_64.whl
triton-3.5.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl
Build Info
- ROCm Version: 7.1
- Python Version: 3.12
- GPU Architectures: gfx942
- Commit: 035f0f1
Installation
These wheels are automatically referenced by the PyPI index at:
https://EmbeddedLLM.github.io/vllm-rocm/simple/
Use: pip install vllm-rocm --index-url https://EmbeddedLLM.github.io/vllm-rocm/simple/
ROCm Wheels - 2025-11-18
Large ML framework wheels for vLLM ROCm
Included Wheels
torch-2.9.0a0+git1c57644-cp312-cp312-linux_x86_64.whl
triton-3.4.0-cp312-cp312-linux_x86_64.whl
Build Info
- ROCm Version: 7.1
- Python Version: 3.12
- GPU Architectures: gfx942
- Commit: 15899db
Installation
These wheels are automatically referenced by the PyPI index at:
https://EmbeddedLLM.github.io/vllm-rocm/simple/
Use: pip install vllm-rocm --index-url https://EmbeddedLLM.github.io/vllm-rocm/simple/
ROCm Wheels - 2025-11-17
Large ML framework wheels for vLLM ROCm
Included Wheels
torch-2.9.0a0+git1c57644-cp312-cp312-linux_x86_64.whl
triton-3.4.0-cp312-cp312-linux_x86_64.whl
Build Info
- ROCm Version: 7.1
- Python Version: 3.12
- GPU Architectures: gfx942
- Commit: 3735664
Installation
These wheels are automatically referenced by the PyPI index at:
https://EmbeddedLLM.github.io/vllm-rocm/simple/
Use: pip install vllm-rocm --index-url https://EmbeddedLLM.github.io/vllm-rocm/simple/
ROCm Wheels - 2025-11-17
Large ML framework wheels for vLLM ROCm
Included Wheels
torch-2.9.0a0+git1c57644-cp312-cp312-linux_x86_64.whl
triton-3.4.0-cp312-cp312-linux_x86_64.whl
Build Info
- ROCm Version: 7.1
- Python Version: 3.12
- GPU Architectures: gfx942
- Commit: 508c45a
Installation
These wheels are automatically referenced by the PyPI index at:
https://EmbeddedLLM.github.io/vllm-rocm/simple/
Use: pip install vllm-rocm --index-url https://EmbeddedLLM.github.io/vllm-rocm/simple/