@@ -48,7 +48,7 @@ $ ./gen_qa_model_repository
4848$ ./gen_qa_custom_ops
4949```
5050
51- This will create multiple model repositories in /tmp/<version >/qa_ *
51+ This will create multiple model repositories in /tmp/\ < version\ > /qa_ *
5252(for example /tmp/24.07/qa_model_repository). The TensorRT models
5353will be created for the GPU on the system that CUDA considers device 0
5454(zero). If you have multiple GPUs on your system see the documentation
@@ -57,14 +57,17 @@ in the scripts for how to target a specific GPU.
5757## Build SDK Image
5858
5959Build the * tritonserver_sdk* image that contains the client
60- libraries, model analyzer, and examples using the following
61- commands. You must first checkout the <client branch > branch of the
62- * client* repo into the clientrepo/ subdirectory. Typically you want to
63- set <client branch > to be the same as your current server branch.
60+ libraries, model analyzer, perf analyzer and examples using the following
61+ commands. You must first checkout the ` <client branch> ` branch of the
62+ * client* repo into the clientrepo/ subdirectory and the ` <perf analyzer branch> `
63+ branch of the * perf_analyzer* repo into the perfanalyzerrepo/ subdirectory
64+ respectively. Typically you want to set both ` <client branch> ` and ` <perf analyzer branch> `
65+ to be the same as your current server branch.
6466
6567```
6668$ cd <server repo root>
6769$ git clone --single-branch --depth=1 -b <client branch> https://github.com/triton-inference-server/client.git clientrepo
70+ $ git clone --single-branch --depth=1 -b <perf analyzer branch> https://github.com/triton-inference-server/perf_analyzer.git perfanalyzerrepo
6871$ docker build -t tritonserver_sdk -f Dockerfile.sdk .
6972```
7073
0 commit comments