Skip to content

Commit a1e6526

Browse files
committed
comma
1 parent fffa0bd commit a1e6526

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

articles/ai-services/openai/how-to/provisioned-throughput-onboarding.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -83,7 +83,7 @@ For example, for `gpt-4.1:2025-04-14`, 1 output token counts as 4 input tokens t
8383
|Global & data zone provisioned scale increment| 5 | 5|5| 5 | 5 |5|5|5|5| 100|100|
8484
|Regional provisioned minimum deployment|25| 50|25| 25 |50 | 25|25|50|25| NA|NA|
8585
|Regional provisioned scale increment|25| 50|25| 25 | 50 | 25|50|50|25|NA|NA|
86-
|Input TPM per PTU|5,400 | 3,000|14,900| 59,400 | 3000 | 2,500|230|2,500|37,000|4,000|4,000|
86+
|Input TPM per PTU|5,400 | 3,000|14,900| 59,400 | 3,000 | 2,500|230|2,500|37,000|4,000|4,000|
8787
|Latency Target Value| 99% > 66 Tokens Per Second\* | 99% > 40 Tokens Per Second\* | 99% > 50 Tokens Per Second\*| 99% > 60 Tokens Per Second\* | 99% > 40 Tokens Per Second\* | 99% > 66 Tokens Per Second\* | 99% > 25 Tokens Per Second\* | 99% > 25 Tokens Per Second\* | 99% > 33 Tokens Per Second\* | 99% > 50 Tokens Per Second\*| 99% > 50 Tokens Per Second\*|
8888

8989
\* Calculated as the average request latency on a per-minute basis across the month.

0 commit comments

Comments
 (0)