Skip to content

Commit 0096b50

Browse files
chzblychEmmaQiaoCh
andauthored
[None][infra] Update upgrade related docs for release 1.2 (#10760) (#10773)
Signed-off-by: qqiao <qqiao@nvidia.com> Signed-off-by: Yanchao Lu <yanchaol@nvidia.com> Co-authored-by: Emma Qiao <qqiao@nvidia.com>
1 parent 7bf4dd9 commit 0096b50

File tree

5 files changed

+16
-11
lines changed

5 files changed

+16
-11
lines changed

README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -11,8 +11,8 @@ This branch is a prototype and not stable for production use. PRs are not accept
1111
[![Documentation](https://img.shields.io/badge/docs-latest-brightgreen.svg?style=flat)](https://nvidia.github.io/TensorRT-LLM/)
1212
[![python](https://img.shields.io/badge/python-3.12-green)](https://www.python.org/downloads/release/python-3123/)
1313
[![python](https://img.shields.io/badge/python-3.10-green)](https://www.python.org/downloads/release/python-31012/)
14-
[![cuda](https://img.shields.io/badge/cuda-13.0.0-green)](https://developer.nvidia.com/cuda-downloads)
15-
[![torch](https://img.shields.io/badge/torch-2.9.0-green)](https://pytorch.org)
14+
[![cuda](https://img.shields.io/badge/cuda-13.1.0-green)](https://developer.nvidia.com/cuda-downloads)
15+
[![torch](https://img.shields.io/badge/torch-2.9.1-green)](https://pytorch.org)
1616
[![version](https://img.shields.io/badge/release-1.3.0rc0-green)](https://github.com/NVIDIA/TensorRT-LLM/blob/main/tensorrt_llm/version.py)
1717
[![license](https://img.shields.io/badge/license-Apache%202-blue)](https://github.com/NVIDIA/TensorRT-LLM/blob/main/LICENSE)
1818

docs/source/installation/linux.md

Lines changed: 6 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -9,15 +9,15 @@
99
Before the pre-built Python wheel can be installed via `pip`, a few
1010
prerequisites must be put into place:
1111

12-
Install CUDA Toolkit 13.0 following the [CUDA Installation Guide for Linux](https://docs.nvidia.com/cuda/cuda-installation-guide-linux/)
12+
Install CUDA Toolkit 13.1 following the [CUDA Installation Guide for Linux](https://docs.nvidia.com/cuda/cuda-installation-guide-linux/)
1313
and make sure `CUDA_HOME` environment variable is properly set.
1414

15-
The `cuda-compat-13-0` package may be required depending on your system's NVIDIA GPU
15+
The `cuda-compat-13-1` package may be required depending on your system's NVIDIA GPU
1616
driver version. For additional information, refer to the [CUDA Forward Compatibility](https://docs.nvidia.com/deploy/cuda-compatibility/forward-compatibility.html).
1717

1818
```bash
1919
# By default, PyTorch CUDA 12.8 package is installed. Install PyTorch CUDA 13.0 package to align with the CUDA version used for building TensorRT LLM wheels.
20-
pip3 install torch==2.9.0 torchvision --index-url https://download.pytorch.org/whl/cu130
20+
pip3 install torch==2.9.1 torchvision --index-url https://download.pytorch.org/whl/cu130
2121

2222
sudo apt-get -y install libopenmpi-dev
2323

@@ -39,6 +39,9 @@
3939
```bash
4040
pip3 install --upgrade pip setuptools && pip3 install tensorrt_llm
4141
```
42+
43+
> **Note:** The TensorRT LLM wheel on PyPI is built with PyTorch 2.9.1. This version may be incompatible with the NVIDIA NGC PyTorch 25.12 container, which uses a more recent PyTorch build from the main branch. If you are using this container or a similar environment, please install the pre-built wheel located at `/app/tensorrt_llm` inside the TensorRT LLM NGC Release container instead.
44+
4245
**This project will download and install additional third-party open source software projects. Review the license terms of these open source projects before use.**
4346

4447
2. Sanity check the installation by running the following in Python (tested on Python 3.12):

docs/source/legacy/reference/support-matrix.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -154,9 +154,9 @@ The following table shows the supported software for TensorRT-LLM.
154154
* -
155155
- Software Compatibility
156156
* - Container
157-
- [25.10](https://docs.nvidia.com/deeplearning/frameworks/support-matrix/index.html)
157+
- [25.12](https://docs.nvidia.com/deeplearning/frameworks/support-matrix/index.html)
158158
* - TensorRT
159-
- [10.13](https://docs.nvidia.com/deeplearning/tensorrt/release-notes/index.html)
159+
- [10.14](https://docs.nvidia.com/deeplearning/tensorrt/release-notes/index.html)
160160
* - Precision
161161
-
162162
- Blackwell (SM100/SM103/SM120) - FP32, FP16, BF16, FP8, FP4, INT8, INT4

jenkins/Build.groovy

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -454,7 +454,7 @@ def runLLMBuild(pipeline, buildFlags, tarName, is_linux_x86_64)
454454
def llmPath = sh (script: "realpath ${LLM_ROOT}",returnStdout: true).trim()
455455
// TODO: Remove after the cmake version is upgraded to 3.31.8
456456
// Get triton tag from docker/dockerfile.multi
457-
def tritonShortTag = "r25.10"
457+
def tritonShortTag = "r25.12"
458458
sh "cd ${LLM_ROOT}/triton_backend/inflight_batcher_llm && mkdir build && cd build && cmake .. -DTRTLLM_DIR=${llmPath} -DTRITON_COMMON_REPO_TAG=${tritonShortTag} -DTRITON_CORE_REPO_TAG=${tritonShortTag} -DTRITON_THIRD_PARTY_REPO_TAG=${tritonShortTag} -DTRITON_BACKEND_REPO_TAG=${tritonShortTag} -DUSE_CXX11_ABI=ON && make -j${buildJobs} install"
459459

460460
// Step 3: packaging wheels into tarfile

jenkins/L0_Test.groovy

Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1988,7 +1988,7 @@ def launchTestListCheck(pipeline)
19881988
def llmPath = sh (script: "realpath .", returnStdout: true).trim()
19891989
def llmSrc = "${llmPath}/TensorRT-LLM/src"
19901990
trtllm_utils.llmExecStepWithRetry(pipeline, script: "pip3 install -r ${llmSrc}/requirements-dev.txt")
1991-
sh "NVIDIA_TRITON_SERVER_VERSION=25.10 LLM_ROOT=${llmSrc} LLM_BACKEND_ROOT=${llmSrc}/triton_backend python3 ${llmSrc}/scripts/check_test_list.py --l0 --qa --waive"
1991+
sh "NVIDIA_TRITON_SERVER_VERSION=25.12 LLM_ROOT=${llmSrc} LLM_BACKEND_ROOT=${llmSrc}/triton_backend python3 ${llmSrc}/scripts/check_test_list.py --l0 --qa --waive"
19921992
} catch (InterruptedException e) {
19931993
throw e
19941994
} catch (Exception e) {
@@ -3431,7 +3431,9 @@ def launchTestJobs(pipeline, testFilter)
34313431
]
34323432

34333433
aarch64SanityCheckConfigs = [
3434-
/* //Disable PY312-UB2404 temporarily since lack of official PyTorch for CUDA 13.1.
3434+
// Workaround PyTorch 2.9.1 vs. 2.10.0a0 incompatibility issue. Once resolved, change back to:
3435+
// 1. DLFW_IMAGE -> UBUNTU_24_04_IMAGE
3436+
// 2. Extra PyTorch CUDA install: false -> true
34353437
"PY312-UB2404": [
34363438
LLM_DOCKER_IMAGE,
34373439
"GH200",
@@ -3440,7 +3442,7 @@ def launchTestJobs(pipeline, testFilter)
34403442
"",
34413443
DLFW_IMAGE,
34423444
false, // Extra PyTorch CUDA 13.0 install
3443-
],*/
3445+
],
34443446
"PY312-DLFW": [
34453447
LLM_DOCKER_IMAGE,
34463448
"GH200",

0 commit comments

Comments
 (0)