Skip to content

Commit 540992d

Browse files
committed
combined 4o models
1 parent d1be6e8 commit 540992d

File tree

1 file changed

+6
-6
lines changed

1 file changed

+6
-6
lines changed

articles/ai-services/openai/concepts/provisioned-throughput.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -44,12 +44,12 @@ Generating output tokens requires more processing and the more tokens generated,
4444

4545
To help with simplifying the sizing effort, the table below outlines the TPM per PTU for the `gpt-4o` and `gpt-4o-mini` models
4646

47-
| | **gpt-4o**, **2024-05-13** | **gpt-4o**, **2024-08-06** | **gpt-4o-mini**, **2024-07-18** |
48-
| --| --| --|--|
49-
| Deployable Increments | 50 | 50 | 25|
50-
| Input TPM per PTU | 2,500 | 2,500 | 37,000 |
51-
| Output TPM per PTU | 833 | 833 | 12,333 |
52-
| Latency target | > 25 tokens per second* | > 25 tokens per second* | > 33 tokens per second* |
47+
| | **gpt-4o**, **2024-05-13** & **gpt-4o**, **2024-08-06** | **gpt-4o-mini**, **2024-07-18** |
48+
| --| --| --|
49+
| Deployable Increments | 50 | 25|
50+
| Input TPM per PTU | 2,500 | 37,000 |
51+
| Output TPM per PTU | 833 | 12,333 |
52+
| Latency target | > 25 tokens per second* | > 33 tokens per second* |
5353

5454
\* Calculated as the average of the per-call average generated tokens on a 1-minute bassis over the month
5555
\** For a full list please see the [AOAI Studio calcualator](https://oai.azure.com/portal/calculator)

0 commit comments

Comments
 (0)