File tree Expand file tree Collapse file tree 7 files changed +11
-11
lines changed
model_analyzer/config/input Expand file tree Collapse file tree 7 files changed +11
-11
lines changed Original file line number Diff line number Diff line change 1212# See the License for the specific language governing permissions and
1313# limitations under the License.
1414
15- ARG BASE_IMAGE=nvcr.io/nvidia/tritonserver:21.09 -py3
16- ARG TRITONSDK_BASE_IMAGE=nvcr.io/nvidia/tritonserver:21.09 -py3-sdk
15+ ARG BASE_IMAGE=nvcr.io/nvidia/tritonserver:21.10 -py3
16+ ARG TRITONSDK_BASE_IMAGE=nvcr.io/nvidia/tritonserver:21.10 -py3-sdk
1717
1818ARG MODEL_ANALYZER_VERSION=1.10.0dev
1919ARG MODEL_ANALYZER_CONTAINER_VERSION=21.11dev
Original file line number Diff line number Diff line change @@ -20,9 +20,9 @@ limitations under the License.
2020
2121** LATEST RELEASE: You are currently on the main branch which tracks
2222under-development progress towards the next release. The latest
23- release of the Triton Model Analyzer is 1.8 .0 and is available on
23+ release of the Triton Model Analyzer is 1.9 .0 and is available on
2424branch
25- [ r21.09 ] ( https://github.com/triton-inference-server/model_analyzer/tree/r21.09 ) .**
25+ [ r21.10 ] ( https://github.com/triton-inference-server/model_analyzer/tree/r21.10 ) .**
2626
2727Triton Model Analyzer is a CLI tool to help with better understanding of the
2828compute and memory requirements of the Triton Inference Server models. These
Original file line number Diff line number Diff line change @@ -119,7 +119,7 @@ profile_models: <comma-delimited-string-list>
119119[ perf_analyzer_max_auto_adjusts: <int> | default: 10 ]
120120
121121# Triton Docker image tag used when launching using Docker mode
122- [ triton_docker_image: <string> | default: nvcr.io/nvidia/tritonserver:21.09 -py3 ]
122+ [ triton_docker_image: <string> | default: nvcr.io/nvidia/tritonserver:21.10 -py3 ]
123123
124124# Triton Server HTTP endpoint url used by Model Analyzer client. Will be ignored if server-launch-mode is not 'remote'".
125125[ triton_http_endpoint: <string> | default: localhost:8000 ]
Original file line number Diff line number Diff line change @@ -26,15 +26,15 @@ Catalog](https://ngc.nvidia.com/catalog/containers/nvidia:tritonserver). You can
2626pull and run the SDK container with the following commands:
2727
2828```
29- $ docker pull nvcr.io/nvidia/tritonserver:21.09 -py3-sdk
29+ $ docker pull nvcr.io/nvidia/tritonserver:21.10 -py3-sdk
3030```
3131
3232If you are not planning to run Model Analyzer with
3333` --triton-launch-mode=docker ` , You can run the SDK container with the following
3434command:
3535
3636```
37- $ docker run -it --gpus all --net=host nvcr.io/nvidia/tritonserver:21.09 -py3-sdk
37+ $ docker run -it --gpus all --net=host nvcr.io/nvidia/tritonserver:21.10 -py3-sdk
3838```
3939
4040You will need to build and install the Triton server binary inside the SDK
@@ -59,7 +59,7 @@ following:
5959$ docker run -it --gpus all \
6060 -v /var/run/docker.sock:/var/run/docker.sock \
6161 -v <path-to-output-model-repo>:<path-to-output-model-repo> \
62- --net=host nvcr.io/nvidia/tritonserver:21.09 -py3-sdk
62+ --net=host nvcr.io/nvidia/tritonserver:21.10 -py3-sdk
6363```
6464
6565Model Analyzer uses ` pdfkit ` for report generation. If you are running Model
Original file line number Diff line number Diff line change @@ -79,7 +79,7 @@ images:
7979
8080 triton:
8181 image: nvcr.io/nvidia/tritonserver
82- tag: 21.09 -py3
82+ tag: 21.10 -py3
8383```
8484
8585The model analyzer executable uses the config file defined in ` helm-chart/templates/config-map.yaml ` . This config can be modified to supply arguments to model analyzer. Only the content under the ` config.yaml ` section of the file should be modified.
Original file line number Diff line number Diff line change @@ -41,4 +41,4 @@ images:
4141
4242 triton :
4343 image : nvcr.io/nvidia/tritonserver
44- tag : 21.09 -py3
44+ tag : 21.10 -py3
Original file line number Diff line number Diff line change 4444DEFAULT_RUN_CONFIG_MAX_PREFERRED_BATCH_SIZE = 16
4545DEFAULT_RUN_CONFIG_PREFERRED_BATCH_SIZE_DISABLE = False
4646DEFAULT_TRITON_LAUNCH_MODE = 'local'
47- DEFAULT_TRITON_DOCKER_IMAGE = 'nvcr.io/nvidia/tritonserver:21.09 -py3'
47+ DEFAULT_TRITON_DOCKER_IMAGE = 'nvcr.io/nvidia/tritonserver:21.10 -py3'
4848DEFAULT_TRITON_HTTP_ENDPOINT = 'localhost:8000'
4949DEFAULT_TRITON_GRPC_ENDPOINT = 'localhost:8001'
5050DEFAULT_TRITON_METRICS_URL = 'http://localhost:8002/metrics'
You can’t perform that action at this time.
0 commit comments