Skip to content

Commit 58fe34d

Browse files
authored
Build: Update main branch post 24.10 release (#7754)
1 parent b9738bb commit 58fe34d

File tree

27 files changed

+65
-62
lines changed

27 files changed

+65
-62
lines changed

Dockerfile.sdk

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@
2929
#
3030

3131
# Base image on the minimum Triton container
32-
ARG BASE_IMAGE=nvcr.io/nvidia/tritonserver:24.09-py3-min
32+
ARG BASE_IMAGE=nvcr.io/nvidia/tritonserver:24.10-py3-min
3333

3434
ARG TRITON_CLIENT_REPO_SUBDIR=clientrepo
3535
ARG TRITON_PA_REPO_SUBDIR=perfanalyzerrepo

README.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -32,8 +32,8 @@
3232

3333
>[!WARNING]
3434
>You are currently on the `main` branch which tracks under-development progress
35-
>towards the next release. The current release is version [2.50.0](https://github.com/triton-inference-server/server/releases/latest)
36-
>and corresponds to the 24.09 container release on NVIDIA GPU Cloud (NGC).
35+
>towards the next release. The current release is version [2.51.0](https://github.com/triton-inference-server/server/releases/latest)
36+
>and corresponds to the 24.10 container release on NVIDIA GPU Cloud (NGC).
3737
3838
Triton Inference Server is an open source inference serving software that
3939
streamlines AI inferencing. Triton enables teams to deploy any AI model from
@@ -91,16 +91,16 @@ Inference Server with the
9191

9292
```bash
9393
# Step 1: Create the example model repository
94-
git clone -b r24.09 https://github.com/triton-inference-server/server.git
94+
git clone -b r24.10 https://github.com/triton-inference-server/server.git
9595
cd server/docs/examples
9696
./fetch_models.sh
9797

9898
# Step 2: Launch triton from the NGC Triton container
99-
docker run --gpus=1 --rm --net=host -v ${PWD}/model_repository:/models nvcr.io/nvidia/tritonserver:24.09-py3 tritonserver --model-repository=/models
99+
docker run --gpus=1 --rm --net=host -v ${PWD}/model_repository:/models nvcr.io/nvidia/tritonserver:24.10-py3 tritonserver --model-repository=/models
100100

101101
# Step 3: Sending an Inference Request
102102
# In a separate console, launch the image_client example from the NGC Triton SDK container
103-
docker run -it --rm --net=host nvcr.io/nvidia/tritonserver:24.09-py3-sdk
103+
docker run -it --rm --net=host nvcr.io/nvidia/tritonserver:24.10-py3-sdk
104104
/workspace/install/bin/image_client -m densenet_onnx -c 3 -s INCEPTION /workspace/images/mug.jpg
105105

106106
# Inference should return the following

TRITON_VERSION

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
2.51.0dev
1+
2.52.0dev

build.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -70,14 +70,14 @@
7070
# incorrectly load the other version of the openvino libraries.
7171
#
7272
TRITON_VERSION_MAP = {
73-
"2.51.0dev": (
74-
"24.10dev", # triton container
75-
"24.09", # upstream container
73+
"2.52.0dev": (
74+
"24.11dev", # triton container
75+
"24.10", # upstream container
7676
"1.19.2", # ORT
77-
"2024.0.0", # ORT OpenVINO
78-
"2024.0.0", # Standalone OpenVINO
77+
"2024.4.0", # ORT OpenVINO
78+
"2024.4.0", # Standalone OpenVINO
7979
"3.2.6", # DCGM version
80-
"0.5.3.post1", # vLLM version
80+
"0.5.5", # vLLM version
8181
"3.12.3", # RHEL Python version
8282
)
8383
}

deploy/aws/values.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
replicaCount: 1
2828

2929
image:
30-
imageName: nvcr.io/nvidia/tritonserver:24.09-py3
30+
imageName: nvcr.io/nvidia/tritonserver:24.10-py3
3131
pullPolicy: IfNotPresent
3232
modelRepositoryPath: s3://triton-inference-server-repository/model_repository
3333
numGpus: 1

deploy/fleetcommand/Chart.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@
2626

2727
apiVersion: v1
2828
# appVersion is the Triton version; update when changing release
29-
appVersion: "2.50.0"
29+
appVersion: "2.51.0"
3030
description: Triton Inference Server (Fleet Command)
3131
name: triton-inference-server
3232
# version is the Chart version; update when changing anything in the chart

deploy/fleetcommand/values.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
replicaCount: 1
2828

2929
image:
30-
imageName: nvcr.io/nvidia/tritonserver:24.09-py3
30+
imageName: nvcr.io/nvidia/tritonserver:24.10-py3
3131
pullPolicy: IfNotPresent
3232
numGpus: 1
3333
serverCommand: tritonserver
@@ -47,13 +47,13 @@ image:
4747
#
4848
# To set model control mode, uncomment and configure below
4949
# TODO: Fix the following url, it is invalid
50-
# See https://github.com/triton-inference-server/server/blob/r24.09/docs/model_management.md
50+
# See https://github.com/triton-inference-server/server/blob/r24.10/docs/model_management.md
5151
# for more details
5252
#- --model-control-mode=explicit|poll|none
5353
#
5454
# Additional server args
5555
#
56-
# see https://github.com/triton-inference-server/server/blob/r24.09/README.md
56+
# see https://github.com/triton-inference-server/server/blob/r24.10/README.md
5757
# for more details
5858

5959
service:

deploy/gcp/values.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
replicaCount: 1
2828

2929
image:
30-
imageName: nvcr.io/nvidia/tritonserver:24.09-py3
30+
imageName: nvcr.io/nvidia/tritonserver:24.10-py3
3131
pullPolicy: IfNotPresent
3232
modelRepositoryPath: gs://triton-inference-server-repository/model_repository
3333
numGpus: 1

deploy/gke-marketplace-app/benchmark/perf-analyzer-script/triton_client.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,7 @@ metadata:
3333
namespace: default
3434
spec:
3535
containers:
36-
- image: nvcr.io/nvidia/tritonserver:24.09-py3-sdk
36+
- image: nvcr.io/nvidia/tritonserver:24.10-py3-sdk
3737
imagePullPolicy: Always
3838
name: nv-triton-client
3939
securityContext:

deploy/gke-marketplace-app/server-deployer/build_and_push.sh

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -27,9 +27,9 @@
2727

2828
export REGISTRY=gcr.io/$(gcloud config get-value project | tr ':' '/')
2929
export APP_NAME=tritonserver
30-
export MAJOR_VERSION=2.50
31-
export MINOR_VERSION=2.50.0
32-
export NGC_VERSION=24.09-py3
30+
export MAJOR_VERSION=2.51
31+
export MINOR_VERSION=2.51.0
32+
export NGC_VERSION=24.10-py3
3333

3434
docker pull nvcr.io/nvidia/$APP_NAME:$NGC_VERSION
3535

0 commit comments

Comments
 (0)