@@ -5,6 +5,7 @@ LLAMA_SERVER_VERSION := latest
55LLAMA_SERVER_VARIANT := cpu
66BASE_IMAGE := ubuntu:24.04
77VLLM_BASE_IMAGE := nvidia/cuda:13.0.2-runtime-ubuntu24.04
8+ VLLM_VERSION ?= 0.12.0
89DOCKER_IMAGE := docker/model-runner:latest
910DOCKER_IMAGE_VLLM := docker/model-runner:latest-vllm-cuda
1011DOCKER_IMAGE_SGLANG := docker/model-runner:latest-sglang
@@ -19,6 +20,7 @@ DOCKER_BUILD_ARGS := \
1920 --build-arg LLAMA_SERVER_VERSION=$(LLAMA_SERVER_VERSION ) \
2021 --build-arg LLAMA_SERVER_VARIANT=$(LLAMA_SERVER_VARIANT ) \
2122 --build-arg BASE_IMAGE=$(BASE_IMAGE ) \
23+ --build-arg VLLM_VERSION='$(VLLM_VERSION ) ' \
2224 --target $(DOCKER_TARGET ) \
2325 -t $(DOCKER_IMAGE )
2426
@@ -232,13 +234,13 @@ vllm-metal-dev:
232234 rm -rf " $( VLLM_METAL_INSTALL_DIR) " ; \
233235 $$ PYTHON_BIN -m venv " $( VLLM_METAL_INSTALL_DIR) " ; \
234236 . " $( VLLM_METAL_INSTALL_DIR) /bin/activate" && \
235- VLLM_VERSION =" 0.13.0" && \
237+ VLLM_UPSTREAM_VERSION =" 0.13.0" && \
236238 WORK_DIR=$$(mktemp -d ) && \
237- curl -fsSL -o " $$ WORK_DIR/vllm.tar.gz" " https://github.com/vllm-project/vllm/releases/download/v$$ VLLM_VERSION /vllm-$$ VLLM_VERSION .tar.gz" && \
239+ curl -fsSL -o " $$ WORK_DIR/vllm.tar.gz" " https://github.com/vllm-project/vllm/releases/download/v$$ VLLM_UPSTREAM_VERSION /vllm-$$ VLLM_UPSTREAM_VERSION .tar.gz" && \
238240 tar -xzf " $$ WORK_DIR/vllm.tar.gz" -C " $$ WORK_DIR" && \
239- pip install -r " $$ WORK_DIR/vllm-$$ VLLM_VERSION /requirements/cpu.txt" && \
241+ pip install -r " $$ WORK_DIR/vllm-$$ VLLM_UPSTREAM_VERSION /requirements/cpu.txt" && \
240242 pip install -e " $( VLLM_METAL_PATH) " && \
241- pip install -r " $$ WORK_DIR/vllm-$$ VLLM_VERSION /requirements/common.txt" && \
243+ pip install -r " $$ WORK_DIR/vllm-$$ VLLM_UPSTREAM_VERSION /requirements/common.txt" && \
242244 rm -rf " $$ WORK_DIR" && \
243245 echo " dev" > " $( VLLM_METAL_INSTALL_DIR) /.vllm-metal-version" ; \
244246 echo " vllm-metal dev installed from $( VLLM_METAL_PATH) "
0 commit comments