Skip to content

Commit 1b8c818

Browse files
dmitry-tokarev-nvkrishung5mc-nvrichardhuo-nv
authored
Prepare r25.05 for merging to main (#8237)
Co-authored-by: Kris Hung <[email protected]> Co-authored-by: Misha Chornyi <[email protected]> Co-authored-by: richardhuo-nv <[email protected]>
1 parent b343cc0 commit 1b8c818

File tree

25 files changed

+47
-43
lines changed

25 files changed

+47
-43
lines changed

Dockerfile.sdk

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@
2929
#
3030

3131
# Base image on the minimum Triton container
32-
ARG BASE_IMAGE=nvcr.io/nvidia/tritonserver:25.04-py3-min
32+
ARG BASE_IMAGE=nvcr.io/nvidia/tritonserver:25.05-py3-min
3333

3434
ARG TRITON_CLIENT_REPO_SUBDIR=clientrepo
3535
ARG TRITON_PA_REPO_SUBDIR=perfanalyzerrepo

README.md

Lines changed: 5 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -29,8 +29,8 @@
2929

3030
>[!WARNING]
3131
>You are currently on the `main` branch which tracks under-development progress
32-
>towards the next release. The current release is version [2.57.0](https://github.com/triton-inference-server/server/releases/latest)
33-
>and corresponds to the 25.04 container release on NVIDIA GPU Cloud (NGC).
32+
>towards the next release. The current release is version [2.58.0](https://github.com/triton-inference-server/server/releases/latest)
33+
>and corresponds to the 25.05 container release on NVIDIA GPU Cloud (NGC).
3434
3535
# Triton Inference Server
3636

@@ -90,16 +90,16 @@ Inference Server with the
9090

9191
```bash
9292
# Step 1: Create the example model repository
93-
git clone -b r25.02 https://github.com/triton-inference-server/server.git
93+
git clone -b r25.05 https://github.com/triton-inference-server/server.git
9494
cd server/docs/examples
9595
./fetch_models.sh
9696

9797
# Step 2: Launch triton from the NGC Triton container
98-
docker run --gpus=1 --rm --net=host -v ${PWD}/model_repository:/models nvcr.io/nvidia/tritonserver:25.02-py3 tritonserver --model-repository=/models --model-control-mode explicit --load-model densenet_onnx
98+
docker run --gpus=1 --rm --net=host -v ${PWD}/model_repository:/models nvcr.io/nvidia/tritonserver:25.05-py3 tritonserver --model-repository=/models --model-control-mode explicit --load-model densenet_onnx
9999

100100
# Step 3: Sending an Inference Request
101101
# In a separate console, launch the image_client example from the NGC Triton SDK container
102-
docker run -it --rm --net=host nvcr.io/nvidia/tritonserver:25.02-py3-sdk /workspace/install/bin/image_client -m densenet_onnx -c 3 -s INCEPTION /workspace/images/mug.jpg
102+
docker run -it --rm --net=host nvcr.io/nvidia/tritonserver:25.05-py3-sdk /workspace/install/bin/image_client -m densenet_onnx -c 3 -s INCEPTION /workspace/images/mug.jpg
103103

104104
# Inference should return the following
105105
Image '/workspace/images/mug.jpg':
@@ -260,4 +260,3 @@ For questions, we recommend posting in our community
260260

261261
Please refer to the [NVIDIA Developer Triton page](https://developer.nvidia.com/nvidia-triton-inference-server)
262262
for more information.
263-

TRITON_VERSION

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
2.58.0dev
1+
2.59.0dev

build.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -73,7 +73,7 @@
7373
DEFAULT_TRITON_VERSION_MAP = {
7474
"release_version": "2.58.0dev",
7575
"triton_container_version": "25.05dev",
76-
"upstream_container_version": "25.04",
76+
"upstream_container_version": "25.05",
7777
"ort_version": "1.22.0",
7878
"ort_openvino_version": "2025.1.0",
7979
"standalone_openvino_version": "2025.1.0",

deploy/aws/values.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
replicaCount: 1
2828

2929
image:
30-
imageName: nvcr.io/nvidia/tritonserver:25.04-py3
30+
imageName: nvcr.io/nvidia/tritonserver:25.05-py3
3131
pullPolicy: IfNotPresent
3232
modelRepositoryPath: s3://triton-inference-server-repository/model_repository
3333
numGpus: 1

deploy/fleetcommand/Chart.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@
2626

2727
apiVersion: v1
2828
# appVersion is the Triton version; update when changing release
29-
appVersion: "2.57.0"
29+
appVersion: "2.58.0"
3030
description: Triton Inference Server (Fleet Command)
3131
name: triton-inference-server
3232
# version is the Chart version; update when changing anything in the chart

deploy/fleetcommand/values.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
replicaCount: 1
2828

2929
image:
30-
imageName: nvcr.io/nvidia/tritonserver:25.04-py3
30+
imageName: nvcr.io/nvidia/tritonserver:25.05-py3
3131
pullPolicy: IfNotPresent
3232
numGpus: 1
3333
serverCommand: tritonserver
@@ -47,13 +47,13 @@ image:
4747
#
4848
# To set model control mode, uncomment and configure below
4949
# TODO: Fix the following url, it is invalid
50-
# See https://github.com/triton-inference-server/server/blob/r25.04/docs/user_guide/model_management.md
50+
# See https://github.com/triton-inference-server/server/blob/r25.05/docs/user_guide/model_management.md
5151
# for more details
5252
#- --model-control-mode=explicit|poll|none
5353
#
5454
# Additional server args
5555
#
56-
# see https://github.com/triton-inference-server/server/blob/r25.04/README.md
56+
# see https://github.com/triton-inference-server/server/blob/r25.05/README.md
5757
# for more details
5858

5959
service:

deploy/gcp/values.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
replicaCount: 1
2828

2929
image:
30-
imageName: nvcr.io/nvidia/tritonserver:25.04-py3
30+
imageName: nvcr.io/nvidia/tritonserver:25.05-py3
3131
pullPolicy: IfNotPresent
3232
modelRepositoryPath: gs://triton-inference-server-repository/model_repository
3333
numGpus: 1

deploy/gke-marketplace-app/benchmark/perf-analyzer-script/triton_client.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,7 @@ metadata:
3333
namespace: default
3434
spec:
3535
containers:
36-
- image: nvcr.io/nvidia/tritonserver:25.04-py3-sdk
36+
- image: nvcr.io/nvidia/tritonserver:25.05-py3-sdk
3737
imagePullPolicy: Always
3838
name: nv-triton-client
3939
securityContext:

deploy/gke-marketplace-app/server-deployer/build_and_push.sh

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -27,9 +27,9 @@
2727

2828
export REGISTRY=gcr.io/$(gcloud config get-value project | tr ':' '/')
2929
export APP_NAME=tritonserver
30-
export MAJOR_VERSION=2.57
31-
export MINOR_VERSION=2.57.0
32-
export NGC_VERSION=25.04-py3
30+
export MAJOR_VERSION=2.58
31+
export MINOR_VERSION=2.58.0
32+
export NGC_VERSION=25.05-py3
3333

3434
docker pull nvcr.io/nvidia/$APP_NAME:$NGC_VERSION
3535

0 commit comments

Comments
 (0)