Skip to content

Commit c61d993

Browse files
authored
Update main post -24.06 (#7380)
1 parent b0b092d commit c61d993

File tree

29 files changed

+72
-83
lines changed

29 files changed

+72
-83
lines changed

Dockerfile.sdk

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@
2929
#
3030

3131
# Base image on the minimum Triton container
32-
ARG BASE_IMAGE=nvcr.io/nvidia/tritonserver:24.05-py3-min
32+
ARG BASE_IMAGE=nvcr.io/nvidia/tritonserver:24.06-py3-min
3333

3434
ARG TRITON_CLIENT_REPO_SUBDIR=clientrepo
3535
ARG TRITON_COMMON_REPO_TAG=main

Dockerfile.win10.min

Lines changed: 17 additions & 28 deletions
Original file line numberDiff line numberDiff line change
@@ -37,9 +37,9 @@ RUN choco install unzip -y
3737
#
3838
# Installing TensorRT
3939
#
40-
ARG TENSORRT_VERSION
41-
ARG TENSORRT_ZIP="TensorRT-${TENSORRT_VERSION}.Windows10.x86_64.cuda-12.0.zip"
42-
ARG TENSORRT_SOURCE=${TENSORRT_ZIP}
40+
ARG TENSORRT_VERSION=10.0.1.6
41+
ARG TENSORRT_ZIP="TensorRT-${TENSORRT_VERSION}.Windows10.x86_64.cuda-12.4.zip"
42+
ARG TENSORRT_SOURCE=https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.0.1/zip/TensorRT-10.0.1.6.Windows10.win10.cuda-12.4.zip
4343
# COPY ${TENSORRT_ZIP} /tmp/${TENSORRT_ZIP}
4444
ADD ${TENSORRT_SOURCE} /tmp/${TENSORRT_ZIP}
4545
RUN unzip /tmp/%TENSORRT_ZIP%
@@ -51,9 +51,9 @@ LABEL TENSORRT_VERSION="${TENSORRT_VERSION}"
5151
#
5252
# Installing cuDNN
5353
#
54-
ARG CUDNN_VERSION
54+
ARG CUDNN_VERSION=9.1.0.70
5555
ARG CUDNN_ZIP=cudnn-windows-x86_64-${CUDNN_VERSION}_cuda12-archive.zip
56-
ARG CUDNN_SOURCE=${CUDNN_ZIP}
56+
ARG CUDNN_SOURCE=https://developer.download.nvidia.com/compute/cudnn/redist/cudnn/windows-x86_64/cudnn-windows-x86_64-9.1.0.70_cuda12-archive.zip
5757
ADD ${CUDNN_SOURCE} /tmp/${CUDNN_ZIP}
5858
RUN unzip /tmp/%CUDNN_ZIP%
5959
RUN move cudnn-* cudnn
@@ -75,7 +75,7 @@ RUN choco install git docker unzip -y
7575
#
7676
# Installing python
7777
#
78-
ARG PYTHON_VERSION=3.8.10
78+
ARG PYTHON_VERSION=3.10.11
7979
ARG PYTHON_SOURCE=https://www.python.org/ftp/python/${PYTHON_VERSION}/python-${PYTHON_VERSION}-amd64.exe
8080
ADD ${PYTHON_SOURCE} python-${PYTHON_VERSION}-amd64.exe
8181
RUN python-%PYTHON_VERSION%-amd64.exe /quiet InstallAllUsers=1 PrependPath=1 Include_doc=0 TargetDir="C:\python%PYTHON_VERSION%"
@@ -88,14 +88,8 @@ LABEL PYTHON_VERSION=${PYTHON_VERSION}
8888
#
8989
# Installing CMake
9090
#
91-
ARG CMAKE_VERSION=3.27.1
92-
ARG CMAKE_FILE=cmake-${CMAKE_VERSION}-windows-x86_64
93-
ARG CMAKE_SOURCE=https://github.com/Kitware/CMake/releases/download/v${CMAKE_VERSION}/${CMAKE_FILE}.zip
94-
95-
ADD ${CMAKE_SOURCE} ${CMAKE_FILE}.zip
96-
RUN unzip %CMAKE_FILE%.zip
97-
RUN move %CMAKE_FILE% "c:\CMake"
98-
RUN setx PATH "c:\CMake\bin;%PATH%"
91+
ARG CMAKE_VERSION=3.29.3
92+
RUN pip install cmake==%CMAKE_VERSION%
9993

10094
ENV CMAKE_TOOLCHAIN_FILE /vcpkg/scripts/buildsystems/vcpkg.cmake
10195
ENV VCPKG_TARGET_TRIPLET x64-windows
@@ -104,27 +98,22 @@ LABEL CMAKE_VERSION=${CMAKE_VERSION}
10498

10599
# Be aware that pip can interact badly with VS cmd shell so need to pip install before
106100
# vsdevcmd.bat (see https://bugs.python.org/issue38989)
107-
108-
109101
#
110102
# Installing Visual Studio BuildTools: VS17 2022
111103
#
112-
ARG BUILDTOOLS_VERSION
104+
ARG BUILDTOOLS_VERSION=17.9.34622.214
113105
# Download collect.exe in case of an install failure.
114106
ADD https://aka.ms/vscollect.exe "C:\tmp\collect.exe"
115107

116108
# Use the latest release channel. For more control, specify the location of an internal layout.
117-
ARG CHANNEL_URL=https://aka.ms/vs/17/release/channel
118-
ADD ${CHANNEL_URL} "C:\tmp\VisualStudio.chman"
119109
# Download the Build Tools bootstrapper.
120-
ARG BUILD_TOOLS_SOURCE=https://aka.ms/vs/17/release/vs_buildtools.exe
110+
# ARG BUILD_TOOLS_SOURCE=https://aka.ms/vs/17/release/vs_buildtools.exe
111+
ARG BUILD_TOOLS_SOURCE=https://download.visualstudio.microsoft.com/download/pr/5e7b923b-7d89-4e14-95b8-a84ab168e243/96b21d216c7954aaf606c6d7ba59a3de991884a8a86c578c767ba349c23188a9/vs_BuildTools.exe
121112
ADD ${BUILD_TOOLS_SOURCE} vs_buildtools.exe
122113
# Install Build Tools with the Microsoft.VisualStudio.Workload.VCTools workload, including recommended.
123114
ARG VS_INSTALL_PATH_WP="C:\BuildTools"
124115
RUN vs_buildtools.exe --quiet --wait --norestart --nocache install \
125116
--installPath %VS_INSTALL_PATH_WP% \
126-
--channelUri "C:\tmp\VisualStudio.chman" \
127-
--installChannelUri "C:\tmp\VisualStudio.chman" \
128117
--add Microsoft.VisualStudio.Workload.VCTools \
129118
--includeRecommended \
130119
--locale "En-us"
@@ -136,15 +125,15 @@ WORKDIR /
136125
#
137126
# Installing Vcpkg
138127
#
139-
ARG VCPGK_VERSION=2023.11.20
128+
ARG VCPGK_VERSION=2024.03.19
140129
RUN git clone --single-branch --depth=1 -b %VCPGK_VERSION% https://github.com/microsoft/vcpkg.git
141130
WORKDIR /vcpkg
142131
RUN bootstrap-vcpkg.bat
143132
RUN vcpkg.exe update
144133
RUN vcpkg.exe install \
145-
b64:x64-windows \
146134
boost-interprocess:x64-windows \
147135
boost-stacktrace:x64-windows \
136+
b64:x64-windows \
148137
openssl-windows:x64-windows \
149138
openssl:x64-windows \
150139
pthread:x64-windows \
@@ -160,8 +149,8 @@ WORKDIR /
160149
# Installing CUDA
161150
#
162151
ARG CUDA_MAJOR=12
163-
ARG CUDA_MINOR=3
164-
ARG CUDA_PATCH=2
152+
ARG CUDA_MINOR=5
153+
ARG CUDA_PATCH=0
165154
ARG CUDA_VERSION=${CUDA_MAJOR}.${CUDA_MINOR}.${CUDA_PATCH}
166155
ARG CUDA_PACKAGES="nvcc_${CUDA_MAJOR}.${CUDA_MINOR} \
167156
cudart_${CUDA_MAJOR}.${CUDA_MINOR} \
@@ -186,15 +175,15 @@ RUN copy "%CUDA_INSTALL_ROOT_WP%\extras\visual_studio_integration\MSBuildExtensi
186175

187176
RUN setx PATH "%CUDA_INSTALL_ROOT_WP%\bin;%PATH%"
188177

189-
ARG CUDNN_VERSION
178+
ARG CUDNN_VERSION=9.1.0.70
190179
ENV CUDNN_VERSION ${CUDNN_VERSION}
191180
COPY --from=dependency_base /cudnn /cudnn
192181
RUN copy cudnn\bin\cudnn*.dll "%CUDA_INSTALL_ROOT_WP%\bin\."
193182
RUN copy cudnn\lib\x64\cudnn*.lib "%CUDA_INSTALL_ROOT_WP%\lib\x64\."
194183
RUN copy cudnn\include\cudnn*.h "%CUDA_INSTALL_ROOT_WP%\include\."
195184
LABEL CUDNN_VERSION="${CUDNN_VERSION}"
196185

197-
ARG TENSORRT_VERSION
186+
ARG TENSORRT_VERSION=10.0.1.6
198187
ENV TRT_VERSION ${TENSORRT_VERSION}
199188
COPY --from=dependency_base /TensorRT /TensorRT
200189
RUN setx PATH "c:\TensorRT\lib;%PATH%"

README.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,7 @@
3333
> [!WARNING]
3434
> ##### LATEST RELEASE
3535
> You are currently on the `main` branch which tracks under-development progress towards the next release.
36-
> The current release is version [2.46.0](https://github.com/triton-inference-server/server/releases/latest) and corresponds to the 24.05 container release on NVIDIA GPU Cloud (NGC).
36+
> The current release is version [2.47.0](https://github.com/triton-inference-server/server/releases/latest) and corresponds to the 24.06 container release on NVIDIA GPU Cloud (NGC).
3737
3838
Triton Inference Server is an open source inference serving software that
3939
streamlines AI inferencing. Triton enables teams to deploy any AI model from
@@ -91,16 +91,16 @@ Inference Server with the
9191

9292
```bash
9393
# Step 1: Create the example model repository
94-
git clone -b r24.05 https://github.com/triton-inference-server/server.git
94+
git clone -b r24.06 https://github.com/triton-inference-server/server.git
9595
cd server/docs/examples
9696
./fetch_models.sh
9797

9898
# Step 2: Launch triton from the NGC Triton container
99-
docker run --gpus=1 --rm --net=host -v ${PWD}/model_repository:/models nvcr.io/nvidia/tritonserver:24.05-py3 tritonserver --model-repository=/models
99+
docker run --gpus=1 --rm --net=host -v ${PWD}/model_repository:/models nvcr.io/nvidia/tritonserver:24.06-py3 tritonserver --model-repository=/models
100100

101101
# Step 3: Sending an Inference Request
102102
# In a separate console, launch the image_client example from the NGC Triton SDK container
103-
docker run -it --rm --net=host nvcr.io/nvidia/tritonserver:24.05-py3-sdk
103+
docker run -it --rm --net=host nvcr.io/nvidia/tritonserver:24.06-py3-sdk
104104
/workspace/install/bin/image_client -m densenet_onnx -c 3 -s INCEPTION /workspace/images/mug.jpg
105105

106106
# Inference should return the following

TRITON_VERSION

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
2.47.0dev
1+
2.48.0dev

build.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -69,10 +69,10 @@
6969
# incorrectly load the other version of the openvino libraries.
7070
#
7171
TRITON_VERSION_MAP = {
72-
"2.47.0dev": (
72+
"2.48.0dev": (
7373
"24.06dev", # triton container
74-
"24.05", # upstream container
75-
"1.18.0", # ORT
74+
"24.06", # upstream container
75+
"1.18.1", # ORT
7676
"2024.0.0", # ORT OpenVINO
7777
"2024.0.0", # Standalone OpenVINO
7878
"3.2.6", # DCGM version

deploy/aws/values.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
replicaCount: 1
2828

2929
image:
30-
imageName: nvcr.io/nvidia/tritonserver:24.05-py3
30+
imageName: nvcr.io/nvidia/tritonserver:24.06-py3
3131
pullPolicy: IfNotPresent
3232
modelRepositoryPath: s3://triton-inference-server-repository/model_repository
3333
numGpus: 1

deploy/fleetcommand/Chart.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@
2626

2727
apiVersion: v1
2828
# appVersion is the Triton version; update when changing release
29-
appVersion: "2.46.0"
29+
appVersion: "2.47.0"
3030
description: Triton Inference Server (Fleet Command)
3131
name: triton-inference-server
3232
# version is the Chart version; update when changing anything in the chart

deploy/fleetcommand/values.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
replicaCount: 1
2828

2929
image:
30-
imageName: nvcr.io/nvidia/tritonserver:24.05-py3
30+
imageName: nvcr.io/nvidia/tritonserver:24.06-py3
3131
pullPolicy: IfNotPresent
3232
numGpus: 1
3333
serverCommand: tritonserver
@@ -47,13 +47,13 @@ image:
4747
#
4848
# To set model control mode, uncomment and configure below
4949
# TODO: Fix the following url, it is invalid
50-
# See https://github.com/triton-inference-server/server/blob/r24.05/docs/model_management.md
50+
# See https://github.com/triton-inference-server/server/blob/r24.06/docs/model_management.md
5151
# for more details
5252
#- --model-control-mode=explicit|poll|none
5353
#
5454
# Additional server args
5555
#
56-
# see https://github.com/triton-inference-server/server/blob/r24.05/README.md
56+
# see https://github.com/triton-inference-server/server/blob/r24.06/README.md
5757
# for more details
5858

5959
service:

deploy/gcp/values.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
replicaCount: 1
2828

2929
image:
30-
imageName: nvcr.io/nvidia/tritonserver:24.05-py3
30+
imageName: nvcr.io/nvidia/tritonserver:24.06-py3
3131
pullPolicy: IfNotPresent
3232
modelRepositoryPath: gs://triton-inference-server-repository/model_repository
3333
numGpus: 1

deploy/gke-marketplace-app/benchmark/perf-analyzer-script/triton_client.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,7 @@ metadata:
3333
namespace: default
3434
spec:
3535
containers:
36-
- image: nvcr.io/nvidia/tritonserver:24.05-py3-sdk
36+
- image: nvcr.io/nvidia/tritonserver:24.06-py3-sdk
3737
imagePullPolicy: Always
3838
name: nv-triton-client
3939
securityContext:

0 commit comments

Comments
 (0)