We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 395a780 commit 7bf7360Copy full SHA for 7bf7360
README.md
@@ -1522,6 +1522,8 @@ to load the model after the server has been started. The model loading API is
1522
currently not supported during the `auto_complete_config` and `finalize`
1523
functions.
1524
1525
+The model loading API applies to repository-managed backends.
1526
+TensorRT-LLM models must be launched via the TensorRT-LLM launcher and cannot be instantiated via pb_utils.load_model(files=...).
1527
## Using BLS with Stateful Models
1528
1529
[Stateful models](https://github.com/triton-inference-server/server/blob/main/docs/user_guide/architecture.md#stateful-models)
0 commit comments