@@ -32,7 +32,6 @@ ENV PYTHON_VERSION=${PYTHON_VERSION}
32
32
RUN microdnf install -y --nodocs \
33
33
python${PYTHON_VERSION}-devel && \
34
34
python${PYTHON_VERSION} -m venv $VIRTUAL_ENV && \
35
- pip install --no-cache -U pip wheel uv && \
36
35
microdnf clean all
37
36
38
37
@@ -113,17 +112,14 @@ COPY examples/*.jinja /app/data/template/
113
112
114
113
# install vllm by running the payload script and then install flashinfer
115
114
116
- ARG VLLM_WHEEL_VERSION
117
- ARG VLLM_WHEEL_INDEX
118
- ARG FLASHINFER_VERSION
115
+ ARG WHEEL_RELEASE
119
116
RUN --mount=type=cache,target=/root/.cache/uv \
120
117
--mount=type=bind,src=payload,target=/workspace/payload \
121
118
--mount=type=secret,id=rhel-ai-private-index-auth/BOT_PAT \
119
+ source ${VIRTUAL_ENV}/bin/activate && \
122
120
env BOT_PAT=$(cat /run/secrets/rhel-ai-private-index-auth/BOT_PAT) \
123
- VLLM_WHEEL_VERSION=${VLLM_VERSION} \
124
- VLLM_WHEEL_INDEX=${VLLM_WHEEL_INDEX} \
125
- ./payload/run.sh && \
126
- uv pip install "${FLASHINFER_VERSION}"
121
+ WHEEL_RELEASE=${WHEEL_RELEASE} \
122
+ ./payload/run.sh
127
123
128
124
ENV HF_HUB_OFFLINE=1 \
129
125
HOME=/home/vllm \
@@ -161,17 +157,15 @@ FROM vllm-openai AS vllm-grpc-adapter
161
157
162
158
USER root
163
159
164
- ARG VLLM_TGIS_ADAPTER_VERSION
160
+ ARG WHEEL_RELEASE
165
161
RUN --mount=type=cache,target=/root/.cache/uv \
166
162
--mount=type=bind,src=payload,target=/workspace/payload \
167
163
--mount=type=secret,id=rhel-ai-private-index-auth/BOT_PAT \
168
164
cd /workspace && \
169
- ls && \
165
+ source ${VIRTUAL_ENV}/bin/activate && \
170
166
env HOME=/root \
171
167
BOT_PAT=$(cat /run/secrets/rhel-ai-private-index-auth/BOT_PAT) \
172
- VLLM_WHEEL_VERSION=${VLLM_VERSION} \
173
- VLLM_TGIS_ADAPTER_VERSION=${VLLM_TGIS_ADAPTER_VERSION} \
174
- VLLM_WHEEL_INDEX=${VLLM_WHEEL_INDEX} \
168
+ WHEEL_RELEASE=${WHEEL_RELEASE} \
175
169
./payload/run.sh
176
170
177
171
0 commit comments