Skip to content
This repository was archived by the owner on Sep 4, 2025. It is now read-only.

Commit a1578c4

Browse files
committed
Dockerfile: use fixed vllm-provided nccl version
1 parent 9543d0b commit a1578c4

File tree

1 file changed

+11
-1
lines changed

1 file changed

+11
-1
lines changed

Dockerfile.ubi

Lines changed: 11 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -250,6 +250,16 @@ RUN --mount=type=bind,from=build,src=/workspace/dist,target=/workspace/dist \
250250
--mount=type=cache,target=/root/.cache/pip \
251251
pip install dist/*.whl --verbose
252252

253+
# vllm requires a specific nccl version built from source distribution
254+
# See https://github.com/NVIDIA/nccl/issues/1234
255+
RUN pip install \
256+
-v \
257+
--force-reinstall \
258+
--no-binary="all" \
259+
--no-cache-dir \
260+
"vllm-nccl-cu12==2.18.1.0.4.0" && \
261+
mv /root/.config/vllm/nccl/cu12/libnccl.so.2.18.1 /opt/vllm/
262+
253263
# Install flash attention (from pre-built wheel)
254264
RUN --mount=type=bind,from=flash-attn-builder,src=/usr/src/flash-attention-v2,target=/usr/src/flash-attention-v2 \
255265
pip install /usr/src/flash-attention-v2/*.whl --no-cache-dir
@@ -267,7 +277,7 @@ ENV HF_HUB_OFFLINE=1 \
267277
PORT=8000 \
268278
GRPC_PORT=8033 \
269279
HOME=/home/vllm \
270-
VLLM_NCCL_SO_PATH=/opt/vllm/lib/python3.11/site-packages/nvidia/nccl/lib/libnccl.so.2 \
280+
VLLM_NCCL_SO_PATH=/opt/vllm/libnccl.so.2.18.1 \
271281
VLLM_USAGE_SOURCE=production-docker-image
272282

273283
# setup non-root user for OpenShift

0 commit comments

Comments
 (0)