Skip to content

Commit d4f8558

Browse files
Merge pull request #266454 from ChrisHMSFT/chrhoder/20240215PTUupdate
[Azure OpenAI] updated statement on concurrency
2 parents 8eff7d6 + 25413d8 commit d4f8558

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

articles/ai-services/openai/concepts/provisioned-throughput.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -115,7 +115,7 @@ We use a variation of the leaky bucket algorithm to maintain utilization below 1
115115

116116
#### How many concurrent calls can I have on my deployment?
117117

118-
The number of concurrent calls you can have at one time is dependent on each call's shape. The service will continue to accept calls until the utilization is above 100%. To determine the approximate number of concurrent calls you can model out the maximum requests per minute for a particular call shape in the [capacity calculator](https://oai.azure.com/portal/calculator). If `max_tokens` is empty, you can assume a value of 1000
118+
The number of concurrent calls you can achieve depends on each call's shape (prompt size, max_token parameter, etc). The service will continue to accept calls until the utilization reach 100%. To determine the approximate number of concurrent calls you can model out the maximum requests per minute for a particular call shape in the [capacity calculator](https://oai.azure.com/portal/calculator). If the system generates less than the number of samplings tokens like max_token, it will accept more requests.
119119

120120
## Next steps
121121

0 commit comments

Comments
 (0)