Skip to content

Commit 6a06234

Browse files
committed
Enable bitsandbytes quantization on warp size 32 AMD GPUs
Signed-off-by: sstamenk <[email protected]>
1 parent 6108946 commit 6a06234

File tree

1 file changed

+3
-0
lines changed

1 file changed

+3
-0
lines changed

vllm/platforms/rocm.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -202,6 +202,9 @@ class RocmPlatform(Platform):
202202
"petit_nvfp4",
203203
"torchao",
204204
]
205+
# bitsandbytes is not supported on GPUs with warp size 64 (gfx9)
206+
if not on_gfx9():
207+
supported_quantization += ["bitsandbytes"]
205208

206209
@classmethod
207210
def get_vit_attn_backend(cls, head_size: int, dtype: torch.dtype) -> "_Backend":

0 commit comments

Comments
 (0)