Skip to content

Commit e70adf2

Browse files
committed
move and update best practices
1 parent a5dfe23 commit e70adf2

File tree

2 files changed

+15
-8
lines changed

2 files changed

+15
-8
lines changed

articles/ai-services/speech-service/batch-transcription-create.md

Lines changed: 0 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -372,14 +372,6 @@ You can store the results of a batch transcription to a writable Azure Blob stor
372372

373373
If you want to store the transcription results in an Azure Blob storage container by using the [Trusted Azure services security mechanism](batch-transcription-audio-data.md#trusted-azure-services-security-mechanism), consider using [Bring-your-own-storage (BYOS)](bring-your-own-storage-speech-resource.md). For more information, see [Use the Bring your own storage (BYOS) Azure AI Foundry resource for speech to text](bring-your-own-storage-speech-resource-speech-to-text.md).
374374

375-
## Bulk submissions and polling
376-
377-
Batch transcription is asynchronous, and requests are processed one at a time in each region. Submitting jobs at a higher rate does not speed up processing. For example, sending 600 or 6,000 requests per minute has no effect on throughput.
378-
379-
[When monitoring job status](./batch-transcription-get.md), polling every few seconds is unnecessary. If you submit multiple jobs, only the first job will be processed initially; subsequent jobs will wait until the first job completes. Polling all jobs frequently increases system load without benefit. Checking status every ten minutes is sufficient, and polling more often than once per minute is not recommended.
380-
381-
To optimize throughput for large-scale batch transcription, consider distributing your jobs across multiple supported Azure regions. This approach can help balance load and reduce overall processing time, provided your data and compliance requirements allow for multi-region usage. Review [region availability](./regions.md) and ensure your storage and resources are accessible from each region you plan to use.
382-
383375
## Related content
384376

385377
- [Learn more about batch transcription](batch-transcription.md)

articles/ai-services/speech-service/batch-transcription.md

Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -34,6 +34,21 @@ To use the batch transcription REST API:
3434
> [!IMPORTANT]
3535
> Batch transcription jobs are scheduled on a best-effort basis. At peak hours it might take up to 30 minutes or longer for a transcription job to start processing. See how to check the current status of a batch transcription job in [this section](batch-transcription-get.md#get-transcription-status).
3636
37+
## Best practices for improving performance
38+
39+
40+
**Request size**: Batch transcription is asynchronous, and requests are processed one at a time in each region. Submitting jobs at a higher rate does not speed up processing. For example, sending 600 or 6,000 requests per minute has no effect on throughput. We recommend submitting ~1000 files in a single `Transcription_Create` request. This way, you send fewer requests overall.
41+
42+
**Time distribution**: Distribute your requests over time: Submit them across several hours rather than sending them all within a few minutes. Backend processing maintains a stable performance level due to fixed bandwidth, so sending requests too quickly doesn’t improve performance.
43+
44+
**Job monitoring**: [When monitoring job status](./batch-transcription-get.md), polling every few seconds is unnecessary. If you submit multiple jobs, only the first job will be processed initially; subsequent jobs will wait until the first job completes. Polling all jobs frequently increases system load without benefit. Checking the status every 10 minutes is sufficient, and polling more often than once per minute is not recommended.
45+
- Because of the sequential processing, you can get job status by checking only a subset of the files: check the first 100 files, and if they're not completed, later batches are likely not competed either. We recommend waiting at least one minute (ideally five minutes) before checking again.
46+
47+
**Avoid peak traffic for API calls**: The `ListFiles`, `Update`, and `Get` API calls behave similarly to the `Create` call and should be minimized during peak traffic times.
48+
49+
**Load balancing**: To optimize throughput for large-scale batch transcription, consider distributing your jobs across multiple supported Azure regions. This approach can help balance load and reduce overall processing time, provided your data and compliance requirements allow for multi-region usage. Review [region availability](./regions.md) and ensure your storage and resources are accessible from each region you plan to use.
50+
51+
3752
## Related content
3853

3954
- [Locate audio files for batch transcription](batch-transcription-audio-data.md)

0 commit comments

Comments
 (0)