Skip to content

Commit 1b2aefe

Browse files
committed
feat: support vLLM nightly builds via wheels.vllm.ai
Signed-off-by: Dorin Geman <dorin.geman@docker.com>
1 parent 88cebee commit 1b2aefe

File tree

3 files changed

+10
-13
lines changed

3 files changed

+10
-13
lines changed

.github/workflows/release.yml

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -340,8 +340,6 @@ jobs:
340340
"LLAMA_SERVER_VARIANT=cuda"
341341
"BASE_IMAGE=nvidia/cuda:13.0.2-runtime-ubuntu24.04"
342342
"VLLM_VERSION=${{ env.VLLM_VERSION }}"
343-
"VLLM_CUDA_VERSION=cu130"
344-
"VLLM_PYTHON_TAG=cp38-abi3"
345343
"VERSION=${{ env.RELEASE_TAG }}"
346344
push: true
347345
sbom: true

Dockerfile

Lines changed: 4 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -106,13 +106,10 @@ USER modelrunner
106106
# Install uv and vLLM as modelrunner user
107107
RUN curl -LsSf https://astral.sh/uv/install.sh | sh \
108108
&& ~/.local/bin/uv venv --python /usr/bin/python3 /opt/vllm-env \
109-
&& if [ "$TARGETARCH" = "amd64" ]; then \
110-
WHEEL_ARCH="manylinux_2_31_x86_64"; \
111-
WHEEL_URL="https://github.com/vllm-project/vllm/releases/download/v${VLLM_VERSION}/vllm-${VLLM_VERSION}%2B${VLLM_CUDA_VERSION}-${VLLM_PYTHON_TAG}-${WHEEL_ARCH}.whl"; \
112-
~/.local/bin/uv pip install --python /opt/vllm-env/bin/python "$WHEEL_URL"; \
113-
else \
114-
~/.local/bin/uv pip install --python /opt/vllm-env/bin/python "vllm==${VLLM_VERSION}"; \
115-
fi
109+
&& printf '%s' "${VLLM_VERSION}" | grep -qE '^(nightly|[0-9]+\.[0-9]+\.[0-9]+|[0-9a-f]{7,40})$' \
110+
|| { echo "Invalid VLLM_VERSION: must be a version (e.g. 0.16.0), 'nightly', or a hex commit hash"; exit 1; } \
111+
&& ~/.local/bin/uv pip install --python /opt/vllm-env/bin/python vllm \
112+
--extra-index-url "https://wheels.vllm.ai/${VLLM_VERSION}/${VLLM_CUDA_VERSION}"
116113

117114
RUN /opt/vllm-env/bin/python -c "import vllm; print(vllm.__version__)" > /opt/vllm-env/version
118115

Makefile

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -5,6 +5,7 @@ LLAMA_SERVER_VERSION := latest
55
LLAMA_SERVER_VARIANT := cpu
66
BASE_IMAGE := ubuntu:24.04
77
VLLM_BASE_IMAGE := nvidia/cuda:13.0.2-runtime-ubuntu24.04
8+
VLLM_VERSION ?= 0.12.0
89
DOCKER_IMAGE := docker/model-runner:latest
910
DOCKER_IMAGE_VLLM := docker/model-runner:latest-vllm-cuda
1011
DOCKER_IMAGE_SGLANG := docker/model-runner:latest-sglang
@@ -19,6 +20,7 @@ DOCKER_BUILD_ARGS := \
1920
--build-arg LLAMA_SERVER_VERSION=$(LLAMA_SERVER_VERSION) \
2021
--build-arg LLAMA_SERVER_VARIANT=$(LLAMA_SERVER_VARIANT) \
2122
--build-arg BASE_IMAGE=$(BASE_IMAGE) \
23+
--build-arg VLLM_VERSION='$(VLLM_VERSION)' \
2224
--target $(DOCKER_TARGET) \
2325
-t $(DOCKER_IMAGE)
2426

@@ -232,13 +234,13 @@ vllm-metal-dev:
232234
rm -rf "$(VLLM_METAL_INSTALL_DIR)"; \
233235
$$PYTHON_BIN -m venv "$(VLLM_METAL_INSTALL_DIR)"; \
234236
. "$(VLLM_METAL_INSTALL_DIR)/bin/activate" && \
235-
VLLM_VERSION="0.13.0" && \
237+
VLLM_UPSTREAM_VERSION="0.13.0" && \
236238
WORK_DIR=$$(mktemp -d) && \
237-
curl -fsSL -o "$$WORK_DIR/vllm.tar.gz" "https://github.com/vllm-project/vllm/releases/download/v$$VLLM_VERSION/vllm-$$VLLM_VERSION.tar.gz" && \
239+
curl -fsSL -o "$$WORK_DIR/vllm.tar.gz" "https://github.com/vllm-project/vllm/releases/download/v$$VLLM_UPSTREAM_VERSION/vllm-$$VLLM_UPSTREAM_VERSION.tar.gz" && \
238240
tar -xzf "$$WORK_DIR/vllm.tar.gz" -C "$$WORK_DIR" && \
239-
pip install -r "$$WORK_DIR/vllm-$$VLLM_VERSION/requirements/cpu.txt" && \
241+
pip install -r "$$WORK_DIR/vllm-$$VLLM_UPSTREAM_VERSION/requirements/cpu.txt" && \
240242
pip install -e "$(VLLM_METAL_PATH)" && \
241-
pip install -r "$$WORK_DIR/vllm-$$VLLM_VERSION/requirements/common.txt" && \
243+
pip install -r "$$WORK_DIR/vllm-$$VLLM_UPSTREAM_VERSION/requirements/common.txt" && \
242244
rm -rf "$$WORK_DIR" && \
243245
echo "dev" > "$(VLLM_METAL_INSTALL_DIR)/.vllm-metal-version"; \
244246
echo "vllm-metal dev installed from $(VLLM_METAL_PATH)"

0 commit comments

Comments
 (0)