Skip to content

Commit 9765b5c

Browse files
authored
[ROCm][Bugfix] Fixed several bugs related to rccl path and attention selector logic (#3699)
1 parent 430530f commit 9765b5c

File tree

4 files changed

+5
-5
lines changed

4 files changed

+5
-5
lines changed

Dockerfile.rocm

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -90,6 +90,6 @@ RUN cd /app \
9090
&& cd ..
9191

9292
RUN python3 -m pip install --upgrade pip
93-
RUN python3 -m pip install --no-cache-dir ray[all]
93+
RUN python3 -m pip install --no-cache-dir ray[all]==2.9.3
9494

9595
CMD ["/bin/bash"]

requirements-rocm.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ starlette
55
requests
66
py-cpuinfo
77
psutil
8-
ray >= 2.9
8+
ray == 2.9.3
99
sentencepiece # Required for LLaMA tokenizer.
1010
numpy
1111
tokenizers>=0.15.0

vllm/attention/backends/xformers.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -405,8 +405,8 @@ def _check_use_naive_attention() -> bool:
405405
if not is_hip():
406406
return False
407407
# For ROCm, check whether flash attention is installed or not.
408-
has_flash_attn = importlib.util.find_spec("flash_attn") is None
409-
if not has_flash_attn:
408+
use_naive_attention = importlib.util.find_spec("flash_attn") is None
409+
if use_naive_attention:
410410
logger.warning("flash_attn is not installed. Using naive attention. "
411411
"This will take significantly more GPU memory.")
412412
return True

vllm/model_executor/parallel_utils/pynccl.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -41,7 +41,7 @@
4141
if torch.version.cuda is not None:
4242
so_file = "libnccl.so.2"
4343
elif torch.version.hip is not None:
44-
so_file = "librccl.so.2"
44+
so_file = "librccl.so.1"
4545
else:
4646
raise ValueError("NCCL only supports CUDA and ROCm backends.")
4747
logger.debug(f"Loading nccl from library {so_file}")

0 commit comments

Comments
 (0)