Skip to content

Conversation

cjluo-nv
Copy link
Collaborator

@cjluo-nv cjluo-nv commented Sep 16, 2025

What does this PR do?

Upgrade TensorRT-LLM 1.1.0RC2 for the following modelopt release.

Summary by CodeRabbit

  • Chores
    • Upgraded the TensorRT-LLM runtime used by the container image to a newer release.
    • Improves compatibility with recent CUDA and driver stacks and streamlines future maintenance.
    • Changelog updated to reflect the dependency upgrade.
    • No changes to application features or user workflows.

@cjluo-nv cjluo-nv requested a review from a team as a code owner September 16, 2025 21:59
@cjluo-nv cjluo-nv requested a review from meenchen September 16, 2025 21:59
Copy link

coderabbitai bot commented Sep 16, 2025

Warning

Rate limit exceeded

@cjluo-nv has exceeded the limit for the number of commits or files that can be reviewed per hour. Please wait 1 minutes and 0 seconds before requesting another review.

⌛ How to resolve this issue?

After the wait time has elapsed, a review can be triggered using the @coderabbitai review command as a PR comment. Alternatively, push new commits to this PR.

We recommend that you space out your commits to avoid hitting the rate limit.

🚦 How do rate limits work?

CodeRabbit enforces hourly rate limits for each developer per organization.

Our paid plans have higher rate limits than the trial, open-source and free plans. In all cases, we re-allow further reviews after a brief timeout.

Please see our FAQ for further information.

📥 Commits

Reviewing files that changed from the base of the PR and between 5c3b19d and 8751fdc.

📒 Files selected for processing (2)
  • CHANGELOG.rst (1 hunks)
  • docker/Dockerfile (1 hunks)

Walkthrough

Updated the Docker base image tag in docker/Dockerfile from nvcr.io/nvidia/tensorrt-llm/release:1.0.0rc6 to nvcr.io/nvidia/tensorrt-llm/release:1.1.0rc2.post2. Also updated CHANGELOG.rst to replace a bug-fix bullet with an entry noting the TensorRT-LLM dependency upgrade. No other changes.

Changes

Cohort / File(s) Summary
Docker base image update
docker/Dockerfile
Bumped TensorRT-LLM base image tag from 1.0.0rc6 to 1.1.0rc2.post2; all other content unchanged.
Changelog edit
CHANGELOG.rst
Replaced the existing bug-fix bullet "Fix attention head ranking logic for pruning Megatron Core GPT models." with "Upgrade TensorRT-LLM dependency to 1.1.0rc2." No other content changed.

Sequence Diagram(s)

(No sequence diagram provided — changes are configuration and documentation edits without control-flow modifications.)

Estimated code review effort

🎯 2 (Simple) | ⏱️ ~10 minutes

Poem

A rabbit sniffed the build log light,
“A newer tag—what pure delight!”
From rc6 we took a hop,
To post2’s patch — then did we stop?
Nibbles of change, a tiny cheer, 🥕🐇

Pre-merge checks and finishing touches

✅ Passed checks (3 passed)
Check name Status Explanation
Description Check ✅ Passed Check skipped - CodeRabbit’s high-level summary is enabled.
Title Check ✅ Passed The title succinctly and accurately summarizes the PR’s primary change: upgrading the TensorRT‑LLM Docker image. It matches the Dockerfile and changelog edits and is concise and clear for reviewers scanning history.
Docstring Coverage ✅ Passed No functions found in the changes. Docstring coverage check skipped.

Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share

Comment @coderabbitai help to get the list of available commands and usage tips.

Copy link

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 0

🧹 Nitpick comments (4)
docker/Dockerfile (4)

3-7: Scope PIP extra index to the commands that need it.

TRT‑LLM no longer requires --extra-index-url to install, so keeping PIP_EXTRA_INDEX_URL globally may unintentionally affect later pip install -r requirements.txt. Prefer passing the flag only where needed. (nvidia.github.io)

Apply:

-ARG PIP_EXTRA_INDEX_URL="https://pypi.nvidia.com"
-ENV PIP_EXTRA_INDEX_URL=$PIP_EXTRA_INDEX_URL \
-    PIP_NO_CACHE_DIR=off \
+ARG PIP_EXTRA_INDEX_URL="https://pypi.nvidia.com"
+ENV PIP_NO_CACHE_DIR=off \
     PIP_CONSTRAINT= \
     TORCH_CUDA_ARCH_LIST="8.0 8.6 8.7 8.9 9.0 10.0+PTX"

Then pass it explicitly where required (see comment on Line 22).


22-23: Pin ModelOpt to a compatible range to avoid surprise upgrades.

Recent TRT‑LLM releases updated the required ModelOpt version; suggest pinning to a compatible minor (e.g., ~=0.27) to keep builds reproducible. (nvidia.github.io)

-RUN pip install -U "nvidia-modelopt[all,dev-test]"
+RUN pip install -U "nvidia-modelopt[all,dev-test]~=0.27" ${PIP_EXTRA_INDEX_URL:+--extra-index-url "$PIP_EXTRA_INDEX_URL"}

9-11: Minor image‑size hardening.

Use --no-install-recommends and clean apt cache to reduce layer size.

-RUN apt-get update && \
-    apt-get install -y libgl1 && \
-    rm -rf /var/lib/apt/lists/*
+RUN apt-get update && \
+    apt-get install -y --no-install-recommends libgl1 && \
+    rm -rf /var/lib/apt/lists/* /var/cache/apt/*

34-34: Avoid 0777; consider a non‑root user.

0777 is overly permissive. Prefer creating a user and chowning the workspace, or at least 0775.

Example:

+# Create non-root user
+RUN useradd -m -u 1000 -s /bin/bash appuser && \
+    mkdir -p /workspace && chown -R appuser:appuser /workspace
-RUN chmod -R 777 /workspace
+USER appuser
📜 Review details

Configuration used: CodeRabbit UI

Review profile: CHILL

Plan: Pro

📥 Commits

Reviewing files that changed from the base of the PR and between 8a5736a and 87d89cf.

📒 Files selected for processing (1)
  • docker/Dockerfile (1 hunks)
⏰ Context from checks skipped due to timeout of 90000ms. You can increase the timeout in your CodeRabbit configuration to a maximum of 15 minutes (900000ms). (3)
  • GitHub Check: linux
  • GitHub Check: build-docs
  • GitHub Check: code-quality
🔇 Additional comments (3)
docker/Dockerfile (3)

15-15: Symlink source may have changed across base tags.

Ensure /app/tensorrt_llm exists in 1.1.0rc2.post2; otherwise the symlink breaks. The verification script in Line 1 checks this.


18-19: Double‑check hard‑coded TRT library/bin paths.

Validate /usr/local/tensorrt/targets/x86_64-linux-gnu/{lib,bin} exist in the new base; these paths have shifted in some images. The Line 1 script covers this.


1-1: Confirm tag availability and decide between rc2 vs rc3.
Docker unavailable here (docker: command not found); verification couldn't complete. Confirm nvcr.io/nvidia/tensorrt-llm/release:1.1.0rc2.post2 exists on NGC and decide whether to upgrade to 1.1.0rc3 or keep rc2.
Location: docker/Dockerfile line 1

@cjluo-nv cjluo-nv enabled auto-merge (squash) September 16, 2025 22:15
Copy link

codecov bot commented Sep 16, 2025

Codecov Report

✅ All modified and coverable lines are covered by tests.
✅ Project coverage is 73.86%. Comparing base (8a5736a) to head (8751fdc).
⚠️ Report is 1 commits behind head on main.

Additional details and impacted files
@@            Coverage Diff             @@
##             main     #327      +/-   ##
==========================================
- Coverage   73.86%   73.86%   -0.01%     
==========================================
  Files         172      172              
  Lines       17409    17409              
==========================================
- Hits        12860    12859       -1     
- Misses       4549     4550       +1     

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

🚀 New features to boost your workflow:
  • ❄️ Test Analytics: Detect flaky tests, report on failures, and find test suite problems.

@cjluo-nv cjluo-nv merged commit 8d0e40f into main Sep 16, 2025
22 checks passed
@cjluo-nv cjluo-nv deleted the cjluo-nv-patch-1 branch September 16, 2025 22:42
yeyu-nvidia pushed a commit that referenced this pull request Sep 18, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants