Skip to content

Commit 525d0b3

Browse files
committed
[Fix] use Flash-Attn-3 wheel in Dockerfile
1 parent 7729439 commit 525d0b3

File tree

1 file changed

+1
-2
lines changed

1 file changed

+1
-2
lines changed

Dockerfile

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -40,8 +40,7 @@ RUN pip install -r /lightllm/requirements.txt --no-cache-dir --ignore-installed
4040

4141
RUN pip install --no-cache-dir nvidia-nccl-cu12==2.25.1 # for allreduce hang issues in multinode H100
4242

43-
RUN git clone https://github.com/Dao-AILab/flash-attention.git -b v2.7.4.post1
44-
RUN cd flash-attention/hopper && MAX_JOBS=1 NVCC_THREADS=1 python setup.py install
43+
RUN pip install --no-cache-dir https://github.com/ModelTC/flash-attn-3-build/releases/download/v2.7.4.post1/flash_attn-3.0.0b1-cp39-cp39-linux_x86_64.whl
4544

4645
COPY . /lightllm
4746
RUN pip install -e /lightllm --no-cache-dir

0 commit comments

Comments
 (0)