We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 42c26bd commit 9fdecf5Copy full SHA for 9fdecf5
tests/models/quantization/test_bitsandbytes.py
@@ -10,13 +10,14 @@
10
11
from tests.quantization.utils import is_quant_method_supported
12
from vllm.platforms import current_platform
13
+from vllm.platforms.rocm import on_gfx9
14
15
from ...utils import compare_two_settings, multi_gpu_test
16
from ..utils import check_embeddings_close, check_logprobs_close
17
18
pytestmark = pytest.mark.skipif(
- current_platform.is_rocm(),
19
- reason="bitsandbytes quantization not supported on ROCm (CUDA-only kernels)",
+ current_platform.is_rocm() and on_gfx9(),
20
+ reason="bitsandbytes quantization not supported on Instinct (warp size 64 limitation)",
21
)
22
23
models_4bit_to_test = [
0 commit comments