Skip to content

Commit 6400c9b

Browse files
Update articles/ai-services/openai/how-to/latency.md
Co-authored-by: Michael <[email protected]>
1 parent bb4c4e8 commit 6400c9b

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

articles/ai-services/openai/how-to/latency.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -61,7 +61,7 @@ Assuming all requests for a given workload are uniform, the prompt tokens and co
6161
##### How to use system level throughput estimates
6262

6363

64-
Once system level throughput has been estimated for a given workload, these estimates can be used to size Standard and Provisioned deployments. For Standard deployments, the input and output TPM values can be combined to estimate the total TPM to be assigned to a given deployment. For Provisioned deployments, the request token usage data (for the dedicated capacity calculator experience) or input and output TPM values (for the deployment capacity calculator experience) can be used to estimate the number of PTUs required to support a given workload.
64+
Once system level throughput has been estimated for a given workload, these estimates can be used to size Standard and Provisioned deployments. For Standard deployments, the input and output TPM values can be combined to estimate the total TPM to be assigned to a given deployment. For Provisioned deployments, the request token usage data or input and output TPM values can be used to estimate the number of PTUs required to support a given workload with the deployment capacity calculator experience.
6565

6666
Here are a few examples for the GPT-4o mini model:
6767

0 commit comments

Comments
 (0)