Skip to content

Commit 3392955

Browse files
Minor Doc updates
Signed-off-by: Keval Morabia <[email protected]>
1 parent 7832b15 commit 3392955

File tree

9 files changed

+54
-26
lines changed

9 files changed

+54
-26
lines changed

README.md

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -75,8 +75,10 @@ pip install -e .[dev]
7575
```
7676

7777
You can also directly use the [TensorRT-LLM docker images](https://catalog.ngc.nvidia.com/orgs/nvidia/teams/tensorrt-llm/containers/release/tags)
78-
(e.g., `nvcr.io/nvidia/tensorrt-llm/release:<version>`),
79-
which have Model Optimizer pre-installed. Visit our [installation guide](https://nvidia.github.io/TensorRT-Model-Optimizer/getting_started/2_installation.html) for more fine-grained control on installed dependencies or for alternative docker images and environment variables to setup.
78+
(e.g., `nvcr.io/nvidia/tensorrt-llm/release:<version>`), which have Model Optimizer pre-installed.
79+
Make sure to upgrade Model Optimizer to the latest version using ``pip`` as described above.
80+
Visit our [installation guide](https://nvidia.github.io/TensorRT-Model-Optimizer/getting_started/2_installation.html) for
81+
more fine-grained control on installed dependencies or for alternative docker images and environment variables to setup.
8082

8183
## Techniques
8284

docs/source/getting_started/_installation_for_Linux.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@ Environment setup
3434
`TensorRT-LLM docker image <https://catalog.ngc.nvidia.com/orgs/nvidia/teams/tensorrt-llm/containers/release/tags>`_,
3535
e.g., ``nvcr.io/nvidia/tensorrt-llm/release:<version>``.
3636

37-
You may upgrade the Model Optimizer to the latest version if not already as described in the next section.
37+
Make sure to upgrade Model Optimizer to the latest version using ``pip`` as described in the next section.
3838

3939
You would also need to setup appropriate environment variables for the TensorRT binaries as follows:
4040

examples/diffusers/README.md

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -27,6 +27,14 @@ Cache Diffusion is a technique that reuses cached outputs from previous diffusio
2727

2828
## Pre-Requisites
2929

30+
### Docker
31+
32+
Please use the TensorRT docker image (e.g., `nvcr.io/nvidia/tensorrt:25.08-py3`) or visit our [installation docs](https://nvidia.github.io/TensorRT-Model-Optimizer/getting_started/2_installation.html) for more information.
33+
34+
Also follow the installation steps below to upgrade to the latest version of Model Optimizer and install example-specific dependencies.
35+
36+
### Local Installation
37+
3038
Install Model Optimizer with `onnx` and `hf` dependencies using `pip` from [PyPI](https://pypi.org/project/nvidia-modelopt/):
3139

3240
```bash

examples/llm_distill/README.md

Lines changed: 11 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -21,15 +21,23 @@ This section focuses on demonstrating how to apply Model Optimizer to perform kn
2121

2222
## Pre-Requisites
2323

24+
### Docker
25+
26+
For Hugging Face models, please use the PyTorch docker image (e.g., `nvcr.io/nvidia/pytorch:25.06-py3`).
27+
For NeMo models, use the NeMo container (e.g., `nvcr.io/nvidia/nemo:25.07`) which has all the dependencies installed.
28+
Visit our [installation docs](https://nvidia.github.io/TensorRT-Model-Optimizer/getting_started/2_installation.html) for more information.
29+
30+
Also follow the installation steps below to upgrade to the latest version of Model Optimizer and install example-specific dependencies.
31+
32+
### Local Installation
33+
2434
For Hugging Face models, install Model Optimizer with `hf` dependencies using `pip` from [PyPI](https://pypi.org/project/nvidia-modelopt/) and install the requirements for the example:
2535

2636
```bash
27-
pip install nvidia-modelopt[hf]
37+
pip install -U nvidia-modelopt[hf]
2838
pip install -r requirements.txt
2939
```
3040

31-
For NeMo models, use the NeMo container `nvcr.io/nvidia/nemo:25.07` or later which has all the dependencies installed.
32-
3341
## Getting Started
3442

3543
### Set up your base models

examples/llm_ptq/README.md

Lines changed: 13 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -25,17 +25,25 @@ This section focuses on Post-training quantization, a technique that reduces mod
2525

2626
## Pre-Requisites
2727

28+
### Docker
29+
30+
For Hugging Face models, please use the TensorRT-LLM docker image (e.g., `nvcr.io/nvidia/tensorrt-llm/release:1.1.0rc2.post2`).
31+
For NeMo models, use the NeMo container (e.g., `nvcr.io/nvidia/nemo:25.07`).
32+
Visit our [installation docs](https://nvidia.github.io/TensorRT-Model-Optimizer/getting_started/2_installation.html) for more information.
33+
34+
Also follow the installation steps below to upgrade to the latest version of Model Optimizer and install example-specific dependencies.
35+
36+
### Local Installation
37+
2838
For Hugging Face models, install Model Optimizer with `hf` dependencies using `pip` from [PyPI](https://pypi.org/project/nvidia-modelopt/) and install the requirements for the example:
2939

3040
```bash
31-
pip install nvidia-modelopt[hf]
41+
pip install -U nvidia-modelopt[hf]
3242
pip install -r requirements.txt
3343
```
3444

35-
If you want to deploy the quantized model on TRT-LLM, you will also need to install the TRT-LLM dependencies as per the [TRT-LLM documentation](https://nvidia.github.io/TensorRT-LLM/quick-start-guide.html#installation).
36-
Visit our [installation docs](https://nvidia.github.io/TensorRT-Model-Optimizer/getting_started/2_installation.html) for more information.
37-
38-
For NeMo models, use the NeMo container `nvcr.io/nvidia/nemo:25.04` or later which has all the dependencies including TRT-LLM installed.
45+
For TensorRT-LLM deployment, please use the TensorRT-LLM docker image or follow their [installation docs](https://nvidia.github.io/TensorRT-LLM/installation/index.html).
46+
Similarly, for vLLM or SGLang deployment, please use their installation docs.
3947

4048
## Getting Started
4149

examples/llm_qat/README.md

Lines changed: 1 addition & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -22,17 +22,7 @@ Quantization Aware Training (QAT) helps to improve the model accuracy beyond pos
2222

2323
## Pre-Requisites
2424

25-
For Hugging Face models, install Model Optimizer with `hf` dependencies using `pip` from [PyPI](https://pypi.org/project/nvidia-modelopt/) and install the requirements for the example:
26-
27-
```bash
28-
pip install nvidia-modelopt[hf]
29-
pip install -r requirements.txt
30-
```
31-
32-
If you want to deploy the quantized model on TRT-LLM, you will also need to install the TRT-LLM dependencies as per the [TRT-LLM documentation](https://nvidia.github.io/TensorRT-LLM/quick-start-guide.html#installation).
33-
Visit our [installation docs](https://nvidia.github.io/TensorRT-Model-Optimizer/getting_started/2_installation.html) for more information.
34-
35-
For NeMo models, use the NeMo container `nvcr.io/nvidia/nemo:25.04` or later which has all the dependencies including TRT-LLM installed.
25+
Please refer to the [llm_ptq/README.md](../llm_ptq/README.md#pre-requisites) for the pre-requisites.
3626

3727
## Getting Started
3828

examples/onnx_ptq/README.md

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -24,14 +24,16 @@ Model Optimizer enables highly performant quantization formats including NVFP4,
2424

2525
### Docker
2626

27-
Please refer to our [Installation Guide](../../README.md#installation) for recommended docker images.
27+
Please use the TensorRT docker image (e.g., `nvcr.io/nvidia/tensorrt:25.08-py3`) or visit our [installation docs](https://nvidia.github.io/TensorRT-Model-Optimizer/getting_started/2_installation.html) for more information.
28+
29+
Also follow the installation steps below to upgrade to the latest version of Model Optimizer and install example-specific dependencies.
2830

2931
### Local Installation
3032

3133
Install Model Optimizer with `onnx` dependencies using `pip` from [PyPI](https://pypi.org/project/nvidia-modelopt/) and install the requirements for the example:
3234

3335
```bash
34-
pip install nvidia-modelopt[onnx]
36+
pip install -U nvidia-modelopt[onnx]
3537
pip install -r requirements.txt
3638
```
3739

examples/pruning/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ This section focuses on applying Model Optimizer's state-of-the-art complementar
2323

2424
## Pre-Requisites
2525

26-
For Minitron pruning for Megatron-LM / NeMo models, use the NeMo container `nvcr.io/nvidia/nemo:25.07` or later which has all the dependencies installed.
26+
For Minitron pruning for Megatron-LM / NeMo models, use the NeMo container (e.g., `nvcr.io/nvidia/nemo:25.07`) which has all the dependencies installed.
2727

2828
For FastNAS pruning for PyTorch Computer Vision models, no additional dependencies are required.
2929

examples/speculative_decoding/README.md

Lines changed: 11 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,16 +25,26 @@ This example focuses on training with Hugging Face. To train with Megatron‑LM,
2525

2626
## Pre-Requisites
2727

28+
### Docker
29+
30+
Please use the PyTorch docker image (e.g., `nvcr.io/nvidia/pytorch:25.06-py3`) or visit our [installation docs](https://nvidia.github.io/TensorRT-Model-Optimizer/getting_started/2_installation.html) for more information.
31+
32+
Also follow the installation steps below to upgrade to the latest version of Model Optimizer and install dataset and example-specific dependencies.
33+
34+
### Local Installation
35+
2836
Install Modelopt with `hf` dependencies and other requirements for this example:
2937

3038
```bash
31-
pip install -e ...
39+
pip install -U nvidia-modelopt[hf]
3240
pip install -r requirements.txt
3341
```
3442

3543
We use [Daring-Anteater](https://huggingface.co/datasets/nvidia/Daring-Anteater) dataset in this example. Download by:
3644

3745
```bash
46+
apt-get update && apt-get install -y git-lfs
47+
git lfs install --system
3848
git clone https://huggingface.co/datasets/nvidia/Daring-Anteater
3949
```
4050

0 commit comments

Comments
 (0)