Skip to content

Commit 0bccda3

Browse files
committed
try it again
1 parent f04066f commit 0bccda3

File tree

1 file changed

+7
-5
lines changed

1 file changed

+7
-5
lines changed

Dockerfile

Lines changed: 7 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -7,20 +7,22 @@ RUN apt-get update -y \
77

88
RUN ldconfig /usr/local/cuda-12.1/compat/
99

10-
# install sglang's dependencies
10+
# update pip
11+
RUN --mount=type=cache,target=/root/.cache/pip python3 -m pip install --upgrade pip
1112

13+
# install sglang's dependencies
1214
# EFRON:
1315
# these guys are unbelivably huge - >80GiB. Took well over ten minutes to install on my machine and used 28GiB(!) of RAM.
1416
# we should consider having a base image with them pre-installed or seeing if we can knock it down a little bit.
1517
RUN --mount=type=cache,target=/root/.cache/pip python3 -m pip install "sglang[all]"
1618
RUN --mount=type=cache,target=/root/.cache/pip python3 -m pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.3
19+
20+
# install our own python dependencies
1721
COPY requirements.txt ./requirements.txt
18-
RUN --mount=type=cache,target=/root/.cache/pip \
19-
python3 -m pip install --upgrade pip && \
20-
python3 -m pip install --upgrade -r app/requirements.txt
22+
RUN --mount-type=cache,target=/root/.cache/pip python3 -m pip install --upgrade -r ./requirements.txt
2123

2224
# not sure why this is here: is a vllm-workspace even in our image?
23-
ENV PYTHONPATH="/:/vllm-workspace"
25+
# ENV PYTHONPATH="/:/vllm-workspace"
2426
COPY ./src/handler.py ./handler.py
2527

2628
# run the serverless worker

0 commit comments

Comments
 (0)