Skip to content

Commit 61af808

Browse files
committed
fied bug on latency amount
1 parent a81ce16 commit 61af808

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

articles/ai-services/openai/concepts/provisioned-throughput.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -50,7 +50,7 @@ To help with simplifying the sizing effort, the table below outlines the TPM per
5050
| Deployable Increments | 50 | 50 | 25|
5151
| Input TPM per PTU | 2,500 | 2,500 | 37,000 |
5252
| Output TPM per PTU | 833 | 833 | 12,333 |
53-
| Latency target | > 25 tokens per second* | > 25 tokens per second* | > 25 tokens per second* |
53+
| Latency target | > 25 tokens per second* | > 25 tokens per second* | > 33 tokens per second* |
5454

5555
\* Calculated as the average of the per-call average generated tokens on a 1-minute bassis over the month
5656
\** For a full list please see the [AOAI Studio calcualator](https://oai.azure.com/portal/calculator)

0 commit comments

Comments
 (0)