You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/applied-ai-services/form-recognizer/concept-general-document.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,13 +1,13 @@
1
1
---
2
2
title: General key-value extraction - Form Recognizer
3
3
titleSuffix: Azure Applied AI Services
4
-
description: Extract key-value paits, tables, selection marks,and text from your documents with Form Recognizer
4
+
description: Extract key-value pairs, tables, selection marks,and text from your documents with Form Recognizer
5
5
author: laujan
6
6
manager: nitinme
7
7
ms.service: applied-ai-services
8
8
ms.subservice: forms-recognizer
9
9
ms.topic: conceptual
10
-
ms.date: 10/14/2022
10
+
ms.date: 11/10/2022
11
11
ms.author: lajanuar
12
12
monikerRange: 'form-recog-3.0.0'
13
13
recommendations: false
@@ -25,7 +25,7 @@ The General document v3.0 model combines powerful Optical Character Recognition
25
25
The general document API supports most form types and will analyze your documents and extract keys and associated values. It's ideal for extracting common key-value pairs from documents. You can use the general document model as an alternative to training a custom model without labels.
26
26
27
27
> [!NOTE]
28
-
> The ```2022-06-30``` and later versions of the general document model add support for selection marks.
28
+
> The ```2022-06-30``` and subsequent versions of the general document model add support for selection marks.
Copy file name to clipboardExpand all lines: articles/applied-ai-services/form-recognizer/concept-id-document.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -28,7 +28,7 @@ The Form Recognizer Identity document (ID) model combines Optical Character Reco
28
28
29
29
## Identity document (ID) processing
30
30
31
-
Identity document (ID) processing involves extraction of data from identity documents whether manually or using OCR based techniques. Examples of identity documents include passports, driver licenses, resident cards, and national identity cards like the social security card in the US. It is an important step in any business process that requires some proof of identity. Examples include customer verification in banks and other financial institutions, mortgage applications, medical visits, claim processing, hospitality industry, and more. Individuals provide some proof of their identity via driver licenses, passports, and other similar documents so that the business can efficiently verify them before providing services and benefits.
31
+
Identity document (ID) processing involves extraction of data from identity documents whether manually or using OCR based techniques. Examples of identity documents include passports, driver licenses, resident cards, and national identity cards like the social security card in the US. It's an important step in any business process that requires some proof of identity. Examples include customer verification in banks and other financial institutions, mortgage applications, medical visits, claim processing, hospitality industry, and more. Individuals provide some proof of their identity via driver licenses, passports, and other similar documents so that the business can efficiently verify them before providing services and benefits.
Copy file name to clipboardExpand all lines: articles/applied-ai-services/form-recognizer/concept-invoice.md
+8-8Lines changed: 8 additions & 8 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -27,7 +27,7 @@ The machine-learning-based invoice model combines powerful Optical Character Rec
27
27
28
28
## Automated invoice processing
29
29
30
-
Automated invoice processing is the process of extracting key accounts payable fields from including invoice line items from invoices and integrating it with your accounts payable (AP) workflows for reviews and payments. Historically, the accounts payable process has been very manual and time consuming. Accurate extraction of key data from invoices is typically the first and one of the most critical steps in the invoice automation process.
30
+
Automated invoice processing is the process of extracting key accounts payable fields from billing account documents. Extracted data includes line items from invoices integrated with your accounts payable (AP) workflows for reviews and payments. Historically, the accounts payable process has been done manually and, hence, very time consuming. Accurate extraction of key data from invoices is typically the first and one of the most critical steps in the invoice automation process.
31
31
32
32
::: moniker range="form-recog-3.0.0"
33
33
@@ -121,7 +121,7 @@ See how data, including customer information, vendor details, and line items, is
121
121
122
122
1. In the **key** field, paste the key you obtained from your Form Recognizer resource.
123
123
124
-
:::image type="content" source="media/fott-select-form-type.png" alt-text="Screenshot: select form type dropdown window.":::
124
+
:::image type="content" source="media/fott-select-form-type.png" alt-text="Screenshot showing the select-form-type dropdown menu.":::
125
125
126
126
1. Select **Run analysis**. The Form Recognizer Sample Labeling tool will call the Analyze Prebuilt API and analyze the document.
127
127
@@ -233,17 +233,17 @@ The Invoice service will extract the text, tables, and 26 invoice fields. Follow
233
233
| VendorName | string | Vendor who has created this invoice | CONTOSO LTD. ||
234
234
| VendorAddress | string | Mailing address for the Vendor | 123 456th St New York, NY, 10001 ||
235
235
| VendorAddressRecipient | string | Name associated with the VendorAddress | Contoso Headquarters ||
236
-
| CustomerAddress | string | Mailing address for the Customer | 123 Other St, Redmond WA, 98052 ||
236
+
| CustomerAddress | string | Mailing address for the Customer | 123 Other Street, Redmond WA, 98052 ||
237
237
| CustomerAddressRecipient | string | Name associated with the CustomerAddress | Microsoft Corp ||
238
-
| BillingAddress | string | Explicit billing address for the customer | 123 Bill St, Redmond WA, 98052 ||
238
+
| BillingAddress | string | Explicit billing address for the customer | 123 Bill Street, Redmond WA, 98052 ||
239
239
| BillingAddressRecipient | string | Name associated with the BillingAddress | Microsoft Services ||
240
-
| ShippingAddress | string | Explicit shipping address for the customer | 123 Ship St, Redmond WA, 98052 ||
240
+
| ShippingAddress | string | Explicit shipping address for the customer | 123 Ship Street, Redmond WA, 98052 ||
241
241
| ShippingAddressRecipient | string | Name associated with the ShippingAddress | Microsoft Delivery ||
242
242
| SubTotal | number | Subtotal field identified on this invoice | $100.00 | 100 |
243
243
| TotalTax | number | Total tax field identified on this invoice | $10.00 | 10 |
244
244
| InvoiceTotal | number | Total new charges associated with this invoice | $110.00 | 110 |
245
245
| AmountDue | number | Total Amount Due to the vendor | $610.00 | 610 |
246
-
| ServiceAddress | string | Explicit service address or property address for the customer | 123 Service St, Redmond WA, 98052 ||
246
+
| ServiceAddress | string | Explicit service address or property address for the customer | 123 Service Street, Redmond WA, 98052 ||
247
247
| ServiceAddressRecipient | string | Name associated with the ServiceAddress | Microsoft Services ||
248
248
| RemittanceAddress | string | Explicit remittance or payment address for the customer | 123 Remit St New York, NY, 10001 ||
249
249
| RemittanceAddressRecipient | string | Name associated with the RemittanceAddress | Contoso Billing ||
@@ -262,7 +262,7 @@ Following are the line items extracted from an invoice in the JSON output respon
262
262
| UnitPrice | number | The net or gross price (depending on the gross invoice setting of the invoice) of one unit of this item | $30.00 | 30 |
263
263
| ProductCode | string| Product code, product number, or SKU associated with the specific line item | A123 ||
264
264
| Unit | string| The unit of the line item, e.g, kg, lb etc. | hours ||
265
-
| Date | date| Date corresponding to each line item. Often it is a date the line item was shipped | 3/4/2021| 2021-03-04 |
265
+
| Date | date| Date corresponding to each line item. Often it's a date the line item was shipped | 3/4/2021| 2021-03-04 |
266
266
| Tax | number | Tax associated with each line item. Possible values include tax amount, tax %, and tax Y/N | 10% ||
267
267
268
268
### JSON output
@@ -271,7 +271,7 @@ The JSON output has three parts:
271
271
272
272
*`"readResults"` node contains all of the recognized text and selection marks. Text is organized by page, then by line, then by individual words.
273
273
*`"pageResults"` node contains the tables and cells extracted with their bounding boxes, confidence, and a reference to the lines and words in "readResults".
274
-
*`"documentResults"` node contains the invoice-specific values and line items that the model discovered. It is where you'll find all the fields from the invoice such as invoice ID, ship to, bill to, customer, total, line items and lots more.
274
+
*`"documentResults"` node contains the invoice-specific values and line items that the model discovered. It's where you'll find all the fields from the invoice such as invoice ID, ship to, bill to, customer, total, line items and lots more.
Copy file name to clipboardExpand all lines: articles/applied-ai-services/form-recognizer/concept-layout.md
+9-21Lines changed: 9 additions & 21 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -26,7 +26,7 @@ The Form Recognizer Layout is an advanced machine-learning based document layout
26
26
27
27
## Document layout analysis
28
28
29
-
Document structure and layout analysis is the process of analyzing a document to extract regions of interest and their inter-relationships. The goal is to extract text and structural elements from the page for building better semantic understanding models. For all extracted text, there are two types of roles that text plays in a document layout. Text, tables, and selection marks are examples of geometric roles. Titles, headings, and footers are examples of logical roles. For example. a reading system requires differentiating text regions from non-textual ones along with their reading order.
29
+
Document structure layout analysis is the process of analyzing a document to extract regions of interest and their inter-relationships. The goal is to extract text and structural elements from the page for building better semantic understanding models. For all extracted text, there are two types of roles that text plays in a document layout. Text, tables, and selection marks are examples of geometric roles. Titles, headings, and footers are examples of logical roles. For example, a reading system requires differentiating text regions from non-textual ones along with their reading order.
30
30
31
31
The following illustration shows the typical components in an image of a sample page.
32
32
@@ -57,7 +57,7 @@ The following illustration shows the typical components in an image of a sample
57
57
58
58
### Data extraction
59
59
60
-
**Starting with v3.0 GA**, it extracts paragraphs and additional structure information like titles, section headings, page header, page footer, page number, and footnote from the document page. These are examples of logical roles described in the previous section. This capability is supported for PDF documents and images (JPG, PNG, BMP, TIFF).
60
+
**Starting with v3.0 GA**, it extracts paragraphs and more structure information like titles, section headings, page header, page footer, page number, and footnote from the document page. These structural elements are examples of logical roles described in the previous section. This capability is supported for PDF documents and images (JPG, PNG, BMP, TIFF).
@@ -99,21 +99,9 @@ The following tools are supported by Form Recognizer v2.1:
99
99
100
100
::: moniker-end
101
101
102
-
Extract data, including text and table structure from documents.
102
+
### Try layout extraction
103
103
104
-
You'll need the following resources:
105
-
106
-
* An Azure subscription—you can [create one for free](https://azure.microsoft.com/free/cognitive-services/)
107
-
108
-
* A [Form Recognizer instance](https://portal.azure.com/#create/Microsoft.CognitiveServicesFormRecognizer) in the Azure portal. You can use the free pricing tier (`F0`) to try the service. After your resource deploys, select **Go to resource** to get your key and endpoint.
109
-
110
-
:::image type="content" source="media/containers/keys-and-endpoint.png" alt-text="Screenshot: keys and endpoint location in the Azure portal.":::
111
-
112
-
:: moniker range="form-recog-3.0.0"
113
-
114
-
### Try Form Recognizer
115
-
116
-
Try extracting data from forms and documents using the Form Recognizer Studio. You'll need the following resources:
104
+
See how data, including text, tables, table headers, selection marks, and structure information is extracted from documents using Form Recognizer. You'll need the following resources:
117
105
118
106
* An Azure subscription—you can [create one for free](https://azure.microsoft.com/free/cognitive-services/)
119
107
@@ -189,7 +177,7 @@ Try extracting data from forms and documents using the Form Recognizer Studio. Y
189
177
190
178
The layout model extracts text, selection marks, tables, paragraphs, and paragraph types (`roles`) from your documents.
191
179
192
-
### Paragraph extraction <sup>🆕</sup>
180
+
### Paragraph extraction
193
181
194
182
The Layout model extracts all identified blocks of text in the `paragraphs` collection as a top level object under `analyzeResults`. Each entry in this collection represents a text block and includes the extracted text as`content`and the bounding `polygon` coordinates. The `span` information points to the text fragment within the top level `content` property that contains the full text from the document.
195
183
@@ -203,7 +191,7 @@ The Layout model extracts all identified blocks of text in the `paragraphs` coll
203
191
]
204
192
```
205
193
206
-
### Paragraph roles<sup> 🆕</sup>
194
+
### Paragraph roles
207
195
208
196
The new machine-learning based page object detection extracts logical roles like titles, section headings, page headers, page footers, and more. The Form Recognizer Layout model assigns certain text blocks in the `paragraphs` collection with their specialized role or type predicted by the model. They're best used with unstructured documents to help understand the layout of the extracted content for a richer semantic analysis. The following paragraph roles are supported:
209
197
@@ -238,7 +226,7 @@ The new machine-learning based page object detection extracts logical roles like
238
226
239
227
### Pages extraction
240
228
241
-
The pages collection is the very first object you see in the service response.
229
+
The pages collection is the first object you see in the service response.
242
230
243
231
```json
244
232
"pages": [
@@ -369,7 +357,7 @@ The second step is to call the [Get Analyze Layout Result](https://westcentralus
369
357
370
358
|Field| Type | Possible values |
371
359
|:-----|:----:|:----|
372
-
|status | string | `notStarted`: The analysis operation has not started.<br /><br />`running`: The analysis operation is in progress.<br /><br />`failed`: The analysis operation has failed.<br /><br />`succeeded`: The analysis operation has succeeded.|
360
+
|status | string | `notStarted`: The analysis operation hasn't started.<br /><br />`running`: The analysis operation is in progress.<br /><br />`failed`: The analysis operation has failed.<br /><br />`succeeded`: The analysis operation has succeeded.|
373
361
374
362
Call this operation iteratively until it returns the `succeeded` value. Use an interval of 3 to 5 seconds to avoid exceeding the requests per second (RPS) rate.
375
363
@@ -399,7 +387,7 @@ Layout API extracts text from documents and images with multiple text angles and
399
387
400
388
### Tables with headers
401
389
402
-
Layout API extracts tables in the `pageResults` section of the JSON output. Documents can be scanned, photographed, or digitized. Tables can be complex with merged cells or columns, with or without borders, and with odd angles. Extracted table information includes the number of columns and rows, row span, and column span. Each cell with its bounding box is output along with information whether it's recognized as part of a header or not. The model predicted header cells can span multiple rows and are not necessarily the first rows in a table. They also work with rotated tables. Each table cell also includes the full text with references to the individual words in the `readResults` section.
390
+
Layout API extracts tables in the `pageResults` section of the JSON output. Documents can be scanned, photographed, or digitized. Tables can be complex with merged cells or columns, with or without borders, and with odd angles. Extracted table information includes the number of columns and rows, row span, and column span. Each cell with its bounding box is output along with information whether it's recognized as part of a header or not. The model predicted header cells can span multiple rows and aren't necessarily the first rows in a table. They also work with rotated tables. Each table cell also includes the full text with references to the individual words in the `readResults` section.
0 commit comments