You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-services/document-intelligence/concept/troubleshoot-latency.md
+64-11Lines changed: 64 additions & 11 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,26 +1,28 @@
1
1
---
2
2
title: Troubleshoot latency issues with Document Intelligence API
3
3
titleSuffix: Azure AI services
4
-
description: Learn troubleshooting tips, remedial solutions, and best practices to address Document Intelligence latency issues.
4
+
description: Learn troubleshooting tips, remedial solutions, and best practices for addressing Document Intelligence latency issues.
5
5
author: laujan
6
6
manager: nitinme
7
7
ms.service: azure-ai-document-intelligence
8
8
ms.topic: troubleshooting
9
-
ms.date: 02/03/2025
9
+
ms.date: 02/05/2025
10
10
ms.author: lajanuar
11
11
---
12
12
13
13
# Troubleshooting latency issues in Azure AI Document Intelligence
14
14
15
-
This article presents troubleshooting tips, remedial solutions, and best practices to address Document Intelligence latency issues. Your applications can encounter latency with using the Document Intelligence service. Latency refers to the duration an API server takes to handle and process an incoming request before delivering the response to the client. The time required to analyze a document varies based on its size (such as the number of pages) and the content on each page. Document Intelligence operates as a multitenant service, ensuring that latency for similar documents is generally comparable, though not always identical. Variability in latency and performance is an inherent characteristic of any microservice-based, stateless, asynchronous service, especially when processing images and large documents on a large scale. Despite continuous efforts to increase hardware capacity and enhance scalability, some latency issues can still arise during runtime.
15
+
This article presents troubleshooting tips, remedial solutions, and best practices to address Document Intelligence latency issues. Latency refers to the duration an API server takes to handle and process an incoming request before delivering the response to the client. The time required to analyze a document varies based on its size (such as the number of pages) and the content on each page. Document Intelligence operates as a multitenant service, ensuring that latency for similar documents is comparable, though not always identical. Variability in latency and performance is an inherent characteristic of any microservice-based, stateless, asynchronous service, especially when processing images and large documents on a large scale. Despite continuous efforts to increase hardware capacity and enhance scalability, some latency issues can still arise during runtime.
16
16
17
17
> [!NOTE]
18
-
> Azure AI services doesn't offer a Service Level Agreement (SLA) for latency.
19
-
> The asynchronous nature of the API allows you to retrieve results for up to 24 hours after the operation is sent to our backend with the request Id returned by the POST operation. If you are unable to retrieve the result within your normal polling sequence, store the request Id and attempt at a different time before retrying. Please refer to our service page for more guidance.
18
+
> Azure AI services don't provide a Service Level Agreement (SLA) for latency.
19
+
> The Document Intelligence API offers asynchronous functionality, allowing you to access results up to 24 hours after sending your request to our backend. Use the request ID provided by the POST operation to retrieve these results. If you encounter issues during your standard polling sequence, save the request ID and try again later before considering a retry. For further assistance, refer to our service page.
20
+
21
+
To evaluate latency, you should first establish baseline metrics for your specific scenario. These metrics give you the expected end-to-end and server latency within the context of your application environment. Once you have these baseline metrics, it becomes easier to distinguish between abnormal and normal conditions.
20
22
21
23
## Check Azure region status
22
24
23
-
If you're experiencing latency issues, the first to check [Azure status](https://azure.status.microsoft/status)to determine whether there is an ongoing outage or issue impacting your services.
25
+
If you're experiencing latency issues, the first step is to check [Azure status](https://azure.status.microsoft/status)for any current outages or issues that might impact your services.
24
26
25
27
* All active events are listed under the `Current Impact` tab.
26
28
@@ -30,11 +32,11 @@ If you're experiencing latency issues, the first to check [Azure status](https:/
30
32
31
33
## Check file size
32
34
33
-
The size of the files you may be sending through the request API. The service parallelizes processing, larger files can lead to longer processing time. Please normalize your measurement as latency per page. Consider raising the issue if you see sustained periods (more than an hour) with latency per page consistently being above 15s.
35
+
Monitor the size of files you send via the request API. Processing larger files in parallel can result in increased processing times. Normalize your metric by measuring latency per page. If you observe sustained periods (exceeding one hour) where latency per page consistently surpasses 15 seconds, consider addressing the issue.
34
36
35
37
## Check Azure Blob storage latency
36
38
37
-
Latency in Azure Storage operations is affected by the size of the request. Larger operations take more time to complete due to the increased volume of data transferred over the network and processed by Azure Storage.
39
+
The size of a request affects latency in Azure Storage operations. Larger operations take more time to complete due to the increased volume of data transferred over the network and processed by Azure Storage.
38
40
39
41
Azure Storage provides two latency metrics for block blobs in the Azure portal:
40
42
@@ -56,9 +58,60 @@ For more information, *see* [Latency in Blob storage](/azure/storage/blobs/stora
56
58
57
59
## Check monitoring metrics for your resource
58
60
59
-
You can monitor performance metrics and set up alerts for your Document Intelligence resource in the Azure portal. To view latency metrics, navigate to your Document Intelligence resource in the Azure portal:
61
+
Azure portal monitors offer insights into your applications to enhance their performance and availability. There are several tools that you can use to monitor your app's performance in the Azure portal:
62
+
63
+
1. On the **Overview** page, select **Monitoring**, select the time period, and review the **Request latency** metrics on page.
64
+
65
+
:::image type="content" source="../media/latency/azure-portal-monitoring.png" alt-text="Screenshot of Azure usage monitoring metrics in the Azure portal.":::
66
+
67
+
1. On the left navigation window, select **Metrics** from the **Monitoring** drop-down menu.
68
+
69
+
* In the main window, select ➕**Add metric**.
70
+
71
+
* Keep the **Scope** and **Metric Namespace** fields unchanged. Add the **Latency** parameter to the **Metric** field and adjust the **Aggregation** field as needed.
72
+
73
+
:::image type="content" source="../media/latency/azure-portal-monitoring-metrics.png" alt-text="Screenshot of add your own metrics setting in the Azure portal.":::
74
+
75
+
## Set a latency alert in the Azure portal
76
+
77
+
Alerts assist you in identifying and resolving issues by providing proactive notifications when Azure Monitor data suggests a potential issue. An alert rule keeps an eye on your data and notifies you when set criteria are met on your specified resource. You can set up an alert in the Azure portal as follows:
78
+
79
+
1. On the left navigation window, select **Alerts** from the **Monitoring** drop-down menu.
80
+
81
+
1. Select the **Create alert rule** button.
82
+
83
+
1. In the new window that opens, select **Latency** from the **Select a signal** drop-down menu.
84
+
85
+
:::image type="content" source="../media/latency/azure-portal-create-alert.png" alt-text="Screenshot of the create an alert rule page in the Azure portal":::
86
+
87
+
1. Configure the alert by completing the fields on the page.
88
+
89
+
1. After you complete the configuration, select **Review ➕ create**
90
+
91
+
92
+
### Contact us
93
+
94
+
If you're unable to resolve long latency issue, [email us](mailto:[email protected]) with the following information:
95
+
96
+
* Model Name
97
+
98
+
* Version
99
+
100
+
* Subscription ID
101
+
102
+
* Resource ID
103
+
104
+
* Timestamp and issue description
105
+
106
+
* Request IDs of the concerning operations (if possible)
107
+
108
+
* Logs
109
+
110
+
* Sample files
111
+
112
+
* JSON file (output/analyze results)
60
113
61
-
*On the **Overview** page, select **Monitoring**, select the time period, and review the **Request latency** metrics on page.
114
+
*Training set (if it's a training issue related to custom neural models)
62
115
63
-
:::image type="content" source="../media/latency/azure-portal-monitoring.png" alt-text="Screenshot of Azure usage monitoring metrics in the Azure portal.":::
64
116
117
+
For more assistance, you can also or use the feedback widget at the bottom of any Microsoft Learn page.
0 commit comments