File tree Expand file tree Collapse file tree 4 files changed +5
-5
lines changed
model_executor/parallel_utils Expand file tree Collapse file tree 4 files changed +5
-5
lines changed Original file line number Diff line number Diff line change @@ -90,6 +90,6 @@ RUN cd /app \
90
90
&& cd ..
91
91
92
92
RUN python3 -m pip install --upgrade pip
93
- RUN python3 -m pip install --no-cache-dir ray[all]
93
+ RUN python3 -m pip install --no-cache-dir ray[all]==2.9.3
94
94
95
95
CMD ["/bin/bash"]
Original file line number Diff line number Diff line change @@ -5,7 +5,7 @@ starlette
5
5
requests
6
6
py-cpuinfo
7
7
psutil
8
- ray >= 2.9
8
+ ray == 2.9.3
9
9
sentencepiece # Required for LLaMA tokenizer.
10
10
numpy
11
11
tokenizers>=0.15.0
Original file line number Diff line number Diff line change @@ -405,8 +405,8 @@ def _check_use_naive_attention() -> bool:
405
405
if not is_hip ():
406
406
return False
407
407
# For ROCm, check whether flash attention is installed or not.
408
- has_flash_attn = importlib .util .find_spec ("flash_attn" ) is None
409
- if not has_flash_attn :
408
+ use_naive_attention = importlib .util .find_spec ("flash_attn" ) is None
409
+ if use_naive_attention :
410
410
logger .warning ("flash_attn is not installed. Using naive attention. "
411
411
"This will take significantly more GPU memory." )
412
412
return True
Original file line number Diff line number Diff line change 41
41
if torch .version .cuda is not None :
42
42
so_file = "libnccl.so.2"
43
43
elif torch .version .hip is not None :
44
- so_file = "librccl.so.2 "
44
+ so_file = "librccl.so.1 "
45
45
else :
46
46
raise ValueError ("NCCL only supports CUDA and ROCm backends." )
47
47
logger .debug (f"Loading nccl from library { so_file } " )
You can’t perform that action at this time.
0 commit comments