|
| 1 | +FROM nvcr.io/nvidia/tritonserver:22.07-py3 |
| 2 | +# NVIDIA Triton |
| 3 | + |
| 4 | +ENV PYTHONUNBUFFERED=1 \ |
| 5 | + _CUDA_COMPAT_PATH="/usr/local/cuda/compat" \ |
| 6 | + PATH="/opt/tritonserver/bin:/usr/local/mpi/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/ucx/bin" \ |
| 7 | + LD_LIBRARY_PATH="/usr/local/cuda/compat/lib:/usr/local/nvidia/lib:/usr/local/nvidia/lib64" \ |
| 8 | + LANG=C.UTF-8 |
| 9 | + |
| 10 | +RUN apt update && \ |
| 11 | + apt install -y --no-install-recommends \ |
| 12 | + ncurses-term \ |
| 13 | + unzip zlib1g-dev htop && \ |
| 14 | + ln -sf /usr/share/terminfo/x/xterm-color /usr/share/terminfo/x/xterm-256color |
| 15 | + |
| 16 | +RUN update-alternatives --install /usr/bin/python python /usr/bin/python3 2 |
| 17 | +COPY ./service-defs /etc/backend.ai/service-defs |
| 18 | +LABEL ai.backend.kernelspec="1" \ |
| 19 | + ai.backend.envs.corecount="OPENBLAS_NUM_THREADS,OMP_NUM_THREADS,NPROC" \ |
| 20 | + ai.backend.features="batch query uid-match user-input" \ |
| 21 | + ai.backend.base-distro="ubuntu16.04" \ |
| 22 | + ai.backend.accelerators="cuda" \ |
| 23 | + ai.backend.resource.min.cpu="1" \ |
| 24 | + ai.backend.resource.min.mem="1g" \ |
| 25 | + ai.backend.resource.min.cuda.device=1 \ |
| 26 | + ai.backend.resource.min.cuda.shares=0.1 \ |
| 27 | + ai.backend.base-distro="ubuntu16.04" \ |
| 28 | + ai.backend.runtime-type="python" \ |
| 29 | + ai.backend.runtime-path="/usr/bin/python" \ |
| 30 | + ai.backend.service-ports="triton-server:preopen:[8000,8001,8002]" |
| 31 | + |
| 32 | +WORKDIR /home/work |
| 33 | +# vim: ft=dockerfile |
0 commit comments