You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-services/document-intelligence/concept-contract.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -31,7 +31,7 @@ The Document Intelligence contract model uses powerful Optical Character Recogni
31
31
32
32
## Automated contract processing
33
33
34
-
Automated contract processing is the process of extracting key contract fields from documents. Historically, the contract analysis process has been done manually and, hence, very time consuming. Accurate extraction of key data from contracts is typically the first and one of the most critical steps in the contract automation process.
34
+
Automated contract processing is the process of extracting key contract fields from documents. Historically, the contract analysis process is achieved manually and, hence, very time consuming. Accurate extraction of key data from contracts is typically the first and one of the most critical steps in the contract automation process.
Copy file name to clipboardExpand all lines: articles/ai-services/document-intelligence/concept-custom.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -62,7 +62,7 @@ The custom template or custom form model relies on a consistent visual template
62
62
63
63
Your training set consists of structured documents where the formatting and layout are static and constant from one document instance to the next. Custom template models support key-value pairs, selection marks, tables, signature fields, and regions. Template models and can be trained on documents in any of the [supported languages](language-support.md). For more information, *see*[custom template models](concept-custom-template.md).
64
64
65
-
If the language of your documents and extraction scenarios supports custom neural models, it's recommended that you use custom neural models over template models for higher accuracy.
65
+
If the language of your documents and extraction scenarios supports custom neural models, we recommend that you use custom neural models over template models for higher accuracy.
66
66
67
67
> [!TIP]
68
68
>
@@ -104,7 +104,7 @@ If the language of your documents and extraction scenarios supports custom neura
104
104
105
105
### Build mode
106
106
107
-
The build custom model operation has added support for the *template* and *neural* custom models. Previous versions of the REST API and SDKs only supported a single build mode that is now known as the *template* mode.
107
+
The build custom model operation adds support for the *template* and *neural* custom models. Previous versions of the REST API and SDKs only supported a single build mode that is now known as the *template* mode.
108
108
109
109
* Template models only accept documents that have the same basic page structure—a uniform visual appearance—or the same relative positioning of elements within the document.
Copy file name to clipboardExpand all lines: articles/ai-services/document-intelligence/concept-general-document.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -23,7 +23,7 @@ ms.author: lajanuar
23
23
24
24
| Feature | version| Model ID |
25
25
|---------- |---------|--------|
26
-
|Layout model with the optional query string parameter **`features=keyValuePairs`** enabled.|• v4:2023-10-31-preview</br>• v3.1:2023-07-31 (GA) |**`prebuilt-layout`**|
26
+
|`Layout` model with the optional query string parameter **`features=keyValuePairs`** enabled.|• v4:2023-10-31-preview</br>• v3.1:2023-07-31 (GA) |**`prebuilt-layout`**|
Copy file name to clipboardExpand all lines: articles/ai-services/document-intelligence/concept-layout.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -163,7 +163,7 @@ See how data, including text, tables, table headers, selection marks, and struct
163
163
164
164
* Select the **Fetch** button.
165
165
166
-
1. Select **Run Layout**. The Document Intelligence Sample Labeling tool calls the Analyze Layout API and analyze the document.
166
+
1. Select **Run Layout**. The Document Intelligence Sample Labeling tool calls the `Analyze Layout` API and analyze the document.
167
167
168
168
:::image type="content" source="media/fott-layout.png" alt-text="Screenshot of `Layout` dropdown window.":::
169
169
@@ -410,7 +410,7 @@ For large multi-page documents, use the `pages` query parameter to indicate spec
410
410
411
411
## The Get Analyze Layout Result operation
412
412
413
-
The second step is to call the [Get Analyze Layout Result](https://westcentralus.dev.cognitive.microsoft.com/docs/services/form-recognizer-api-v2-1/operations/GetAnalyzeLayoutResult) operation. This operation takes as input the Result ID the Analyze Layout operation created. It returns a JSON response that contains a **status** field with the following possible values.
413
+
The second step is to call the [Get Analyze Layout Result](https://westcentralus.dev.cognitive.microsoft.com/docs/services/form-recognizer-api-v2-1/operations/GetAnalyzeLayoutResult) operation. This operation takes as input the Result ID the `Analyze Layout` operation created. It returns a JSON response that contains a **status** field with the following possible values.
Copy file name to clipboardExpand all lines: articles/ai-services/document-intelligence/faq.yml
+10-10Lines changed: 10 additions & 10 deletions
Original file line number
Diff line number
Diff line change
@@ -129,11 +129,11 @@ sections:
129
129
What is the confidence score and how is it calculated?
130
130
answer: |
131
131
132
-
A confidence score indicates probability by measuring the degree of statistical certainty that the extracted result has been detected correctly.
132
+
A confidence score indicates probability by measuring the degree of statistical certainty that the extracted result is detected correctly.
133
133
134
134
The confidence value range is a percentage between 0% (low) and 100% (high).
135
135
136
-
It's best to target a score of 80% or higher. For more sensitive cases, like financial or medical records, a score of close to 100% is recommended. You may also require human review.
136
+
It's best to target a score of 80% or higher. For more sensitive cases, like financial or medical records, a score of close to 100% is recommended. You can also require human review.
137
137
138
138
See [Interpret and improve accuracy and confidence scores](concept-accuracy-confidence.md#confidence-scores)
139
139
@@ -159,7 +159,7 @@ sections:
159
159
What is a bounding box?
160
160
answer: |
161
161
162
-
A bounding box is an abstract rectangle that surrounds text elements on a document or form. It's used as a reference point for object detection.
162
+
A bounding box is an abstract rectangle that surrounds text elements on a document or form and is used as a reference point for object detection.
163
163
164
164
- The bounding box specifies position using an x and y coordinate plane presented in an array of four numerical pairs. Each pair represents a corner of the box in the following order: top-left, top-right, bottom-right, bottom-left.
165
165
@@ -277,7 +277,7 @@ sections:
277
277
278
278
- When analyzing Microsoft Word and HTML files supported by only the Read model, pages are counted in blocks of 3,000 characters each. For example, if your document contains 7,000 characters, the two pages with 3,000 characters each and one page with 1,000 characters adds up to a total of three pages.
279
279
280
-
- In addition, when using the Read model, if your Microsoft Word, Excel, and PowerPoint pages have embedded images, each image is analyzed and counted as a page. Therefore, the total analyzed pages for Microsoft Office documents are equal to the sum of total text pages and total images analyzed. In the previous example if the document contains two embedded images, the total page count in the service output is three text pages plus two images equaling a total of five pages.
280
+
- When using the Read model, if your Microsoft Word, Excel, and PowerPoint pages with embedded images, each image is analyzed and counted as a page. Therefore, the total analyzed pages for Microsoft Office documents are equal to the sum of total text pages and total images analyzed. In the previous example if the document contains two embedded images, the total page count in the service output is three text pages plus two images equaling a total of five pages.
281
281
282
282
- Training a custom model is always free with Document Intelligence. You’re only charged when a model is used to analyze a document.
283
283
@@ -314,7 +314,7 @@ sections:
314
314
answer: |
315
315
Document Intelligence is a multi-tenanted service where latency for similar documents is comparable but not always identical. The time to analyze a document depends on the size (for example, number of pages) and associated content on each page.
316
316
317
-
Latency is the amount of time it takes for an API server to handle and process an incoming request and deliver the outgoing response to the client. Occasional variability in latency and performance is inherent in any micro-service-based, stateless, asynchronous service that processes images and large documents at scale. While we're continuously scaling up the hardware and capacity and scaling capabilities, you may still see latency issues at run time.
317
+
Latency is the amount of time it takes for an API server to handle and process an incoming request and deliver the outgoing response to the client. Occasional variability in latency and performance is inherent in any micro-service-based, stateless, asynchronous service that processes images and large documents at scale. While we're continuously scaling up the hardware and capacity and scaling capabilities, you can still see latency issues at run time.
318
318
319
319
- name: Custom models
320
320
questions:
@@ -367,7 +367,7 @@ sections:
367
367
368
368
- Model Compose is currently available only for custom models trained with labels.
369
369
370
-
- Analyzing a document with composed models is identical to analyzing a document with a single model, the analyze result returns a ```docType``` property indicating which of the component models was selected for analyzing the document. There's no change in pricing for analyzing a document with an individual custom model or a composed custom model.
370
+
- Analyzing a document with composed models is identical to analyzing a document with a single model, the Analyze result returns a ```docType``` property indicating which of the component models was selected for analyzing the document. There's no change in pricing for analyzing a document with an individual custom model or a composed custom model.
371
371
372
372
Learn more about [composed models](concept-custom.md).
373
373
@@ -431,11 +431,11 @@ sections:
431
431
Consider setting a longer duration period for the time you're using your storage account with Document Intelligence.
432
432
433
433
- question: |
434
-
If my storage account is behind a VNet or firewall, how do I give Document Intelligence access to my storage account data?
434
+
If my storage account is behind a virtual network or firewall, how do I give Document Intelligence access to my storage account data?
435
435
answer: |
436
-
If you have an Azure storage account protected by a Virtual Network (VNet) or firewall, Document Intelligence can’t directly access your storage account. However, Private Azure storage account access and authentication support [managed identities for Azure resources](../../active-directory/managed-identities-azure-resources/overview.md). Once a managed identity is enabled, the Document Intelligence service can access your storage account using an assigned managed identity credential.
436
+
If you have an Azure storage account protected by a Virtual Network (virtual network) or firewall, Document Intelligence can’t directly access your storage account. However, Private Azure storage account access and authentication support [managed identities for Azure resources](../../active-directory/managed-identities-azure-resources/overview.md). Once a managed identity is enabled, the Document Intelligence service can access your storage account using an assigned managed identity credential.
437
437
438
-
If you intend to analyze your private storage account data with FOTT, the tool must be deployed behind the VNet or firewall.
438
+
If you intend to analyze your private storage account data with FOTT, the tool must be deployed behind the `virtual network` or firewall.
439
439
440
440
Learn how to [create and use a managed identity for your Document Intelligence resource](managed-identities.md)
441
441
@@ -471,7 +471,7 @@ sections:
471
471
Why am I receiving an AuthorizationFailure error on Project Sharing, Auto Label, or OCR Upgrade when my Document Intelligence or Storage Account resource is configured with a firewall?
472
472
answer: |
473
473
474
-
Add our website IP address, 20.3.165.95, to the firewall allowlist for both Document Intelligence and Storage Account resources. This is Document Intelligence Studio's dedicated IP address and can be safely allowed.
474
+
Add our website IP address, 20.3.165.95, to the firewall allowlist for both Document Intelligence and Storage Account resources. This unique address is Document Intelligence Studio's dedicated IP address and can be safely allowed.
475
475
476
476
- question: |
477
477
Can I reuse or customize the labeling experience from Studio and build it into my own application?
To enable add-on capabilities, use the `features` query parameter in the POST request. There are four add-on capabilities available with the 2023-07-31 (GA) release: *ocr.highResolution*, *ocr.formula*, *ocr.font*, and *queryFields.premium*. To learn more about each of the capabilities, see [Custom models](../../../concept-accuracy-confidence.md).
71
+
To enable add-on capabilities, use the `features` query parameter in the POST request. There are four add-on capabilities available with the `2023-07-31` (GA) release: *ocr.highResolution*, *ocr.formula*, *ocr.font*, and *queryFields.premium*. To learn more about each of the capabilities, see [Custom models](../../../concept-accuracy-confidence.md).
72
72
73
73
You can only call the *highResolution*, *formula*, and *font* capabilities for the Read and Layout model, and the *queryFields* capability for the General Documents model. The following example shows how to call the *highResolution*, *formula*, and *font* capabilities for the Layout model.
74
74
@@ -87,7 +87,7 @@ You receive a `202 (Success)` response that includes an `Operation-location` hea
87
87
88
88
### Get analyze result (GET Request)
89
89
90
-
After you call the [Analyze document](/rest/api/aiservices/document-models/analyze-document?view=rest-aiservices-2023-07-31&preserve-view=true&tabs=HTTP) API, call the [Get analyze result}(/rest/api/aiservices/document-models/get-analyze-result?view=rest-aiservices-2023-07-31&preserve-view=true&tabs=HTTP) API to get the status of the operation and the extracted data.
90
+
After you call the [Analyze document](/rest/api/aiservices/document-models/analyze-document?view=rest-aiservices-2023-07-31&preserve-view=true&tabs=HTTP) API, call the [`Get analyze` result}(/rest/api/aiservices/document-models/get-analyze-result?view=rest-aiservices-2023-07-31&preserve-view=true&tabs=HTTP) API to get the status of the operation and the extracted data.
91
91
92
92
<!-- markdownlint-disable MD024 -->
93
93
@@ -127,9 +127,9 @@ The *json_pp* command tool ships with macOS and can be used as a JSON formatter
127
127
128
128
#### [Linux](#tab/linux)
129
129
130
-
The *json_pp* command line tool is preinstalled in most Linux distributions. If it's not included, you can use your distribution's package manager to install it.
130
+
The *json_pp* command line tool is preinstalled in most Linux distributions. If it isn't included, you can use your distribution's package manager to install it.
131
131
132
-
- Pretty print the JSON output by including `| json_pp` with your GET requests.
132
+
- Pretty print the JSON output by including `| json_pp` with your `GET` requests.
133
133
134
134
```console
135
135
curl -i -X GET "<endpoint>formrecognizer/documentModels/prebuilt-read/analyzeResults/0e49604a-2d8e-4b15-b6b8-bb456e5d3e0a?api-version=2023-07-31"-H "Ocp-Apim-Subscription-Key: <subscription key>" | json_pp
You receive a `200 (Success)` response with JSON output. The first field, `status`, indicates the status of the operation. If the operation isn't complete, the value of `status` is `running` or `notStarted`. Call the API again, either manually or through a script. We recommend an interval of one second or more between calls.
161
161
162
-
Visit the Azure samples repository on GitHub to view the GET response for each of the Document Intelligence models:
162
+
Visit the Azure samples repository on GitHub to view the `GET` response for each of the Document Intelligence models:
0 commit comments