Skip to content

Commit d646e96

Browse files
committed
resolve markers
1 parent c10a5d8 commit d646e96

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

articles/ai-services/openai/concepts/provisioned-throughput.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@ The amount of throughput (tokens per minute or TPM) a deployment gets per PTU is
4343

4444
To help with simplifying the sizing effort, the following table outlines the TPM per PTU for the `gpt-4o` and `gpt-4o-mini` models which represent the max TPM assuming all traffic is either input or output. To understand how different ratios of input and output tokens impact your Max TPM per PTU, see the [Azure OpenAI capacity calculator](https://oai.azure.com/portal/calculator). The table also shows Service Level Agreement (SLA) Latency Target Values per model. For more information about the SLA for Azure OpenAI Service, see the [Service Level Agreements (SLA) for Online Services page](https://www.microsoft.com/licensing/docs/view/Service-Level-Agreements-SLA-for-Online-Services?lang=1)
4545

46-
| **gpt-4o**, **2024-05-13** & **gpt-4o**, **2024-08-06** | **gpt-4o-mini**, **2024-07-18** |
46+
|| **gpt-4o**, **2024-05-13** & **gpt-4o**, **2024-08-06** | **gpt-4o-mini**, **2024-07-18** |
4747
| --- | --- | --- |
4848
|Global provisioned minimum deployment|15|15|
4949
|Global provisioned scale increment|5|5|
@@ -195,4 +195,4 @@ The number of concurrent calls you can achieve depends on each call's shape (pro
195195
## Next steps
196196

197197
- [Learn about the onboarding steps for provisioned deployments](../how-to/provisioned-throughput-onboarding.md)
198-
- [Provisioned Throughput Units (PTU) getting started guide](../how-to//provisioned-get-started.md)
198+
- [Provisioned Throughput Units (PTU) getting started guide](../how-to//provisioned-get-started.md)

0 commit comments

Comments
 (0)