Skip to content

Commit a368b48

Browse files
committed
updated title
1 parent 540992d commit a368b48

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

articles/ai-services/openai/concepts/provisioned-throughput.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ An Azure OpenAI Deployment is a unit of management for a specific OpenAI Model.
3737
| Estimating size | Provided calculator in the studio & benchmarking script. |
3838

3939

40-
## How much thoughput you get for each model
40+
## How much thoughput per PTU you get for each model
4141
The amount of throughput (tokens per minute or TPM) a deployment gets per PTU is a function of the input and output tokens being generated.
4242

4343
Generating output tokens requires more processing and the more tokens generated, the lower the overall TPM per PTU. Provisioned deployments dynamically balance the two, so users do not have to set specific input and output limits. This means the service is resilient to fluctuations in the workload shape.

0 commit comments

Comments
 (0)