Skip to content

Commit 22c120e

Browse files
authored
[None][doc] Update docker cmd in quick start guide and trtllm-serve … (#7787)
Signed-off-by: nv-guomingz <[email protected]>
1 parent a838c2c commit 22c120e

File tree

4 files changed

+30
-3
lines changed

4 files changed

+30
-3
lines changed

docs/source/commands/trtllm-serve/run-benchmark-with-trtllm-serve.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,9 +24,10 @@ TensorRT LLM distributes the pre-built container on [NGC Catalog](https://catalo
2424
You can launch the container using the following command:
2525

2626
```bash
27-
docker run --rm --ipc host -p 8000:8000 --gpus all -it nvcr.io/nvidia/tensorrt-llm/release
27+
docker run --rm -it --ipc host -p 8000:8000 --gpus all --ulimit memlock=-1 --ulimit stack=67108864 nvcr.io/nvidia/tensorrt-llm/release:x.y.z
2828
```
2929

30+
3031
## Start the trtllm-serve service
3132
> [!WARNING]
3233
> The commands and configurations presented in this document are for illustrative purposes only.

docs/source/conf.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -164,7 +164,7 @@ def tag_role(name, rawtext, text, lineno, inliner, options=None, content=None):
164164

165165

166166
def setup(app):
167-
from helper import generate_examples, generate_llmapi
167+
from helper import generate_examples, generate_llmapi, update_version
168168

169169
from tensorrt_llm.llmapi.utils import tag_llm_params
170170
tag_llm_params()
@@ -173,6 +173,7 @@ def setup(app):
173173

174174
generate_examples()
175175
generate_llmapi()
176+
update_version()
176177

177178

178179
def gen_cpp_doc(ofile_name: str, header_dir: str, summary: str):

docs/source/helper.py

Lines changed: 25 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,6 @@
1+
import importlib.util
12
import logging
3+
import os
24
import re
35
from dataclasses import dataclass
46
from itertools import chain, groupby
@@ -340,6 +342,29 @@ def generate_llmapi():
340342
f.write(content)
341343

342344

345+
def update_version():
346+
version_path = os.path.abspath(
347+
os.path.join(os.path.dirname(__file__),
348+
"../../tensorrt_llm/version.py"))
349+
spec = importlib.util.spec_from_file_location("version_module",
350+
version_path)
351+
version_module = importlib.util.module_from_spec(spec)
352+
spec.loader.exec_module(version_module)
353+
version = version_module.__version__
354+
file_list = [
355+
"docs/source/quick-start-guide.md",
356+
"docs/source/commands/trtllm-serve/run-benchmark-with-trtllm-serve.md"
357+
]
358+
for file in file_list:
359+
file_path = os.path.abspath(
360+
os.path.join(os.path.dirname(__file__), "../../" + file))
361+
with open(file_path, "r") as f:
362+
content = f.read()
363+
content = content.replace("x.y.z", version)
364+
with open(file_path, "w") as f:
365+
f.write(content)
366+
367+
343368
if __name__ == "__main__":
344369
import os
345370
path = os.environ["TEKIT_ROOT"] + "/examples/llm-api/llm_inference.py"

docs/source/quick-start-guide.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ This is the starting point to try out TensorRT LLM. Specifically, this Quick Sta
88
## Launch Docker on a node with NVIDIA GPUs deployed
99

1010
```bash
11-
docker run --ipc host --gpus all -p 8000:8000 -it nvcr.io/nvidia/tensorrt-llm/release
11+
docker run --rm -it --ipc host --gpus all --ulimit memlock=-1 --ulimit stack=67108864 -p 8000:8000 nvcr.io/nvidia/tensorrt-llm/release:x.y.z
1212
```
1313

1414

0 commit comments

Comments
 (0)