You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/applied-ai-services/form-recognizer/concept-business-card.md
+4-4Lines changed: 4 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -7,7 +7,7 @@ manager: nitinme
7
7
ms.service: applied-ai-services
8
8
ms.subservice: forms-recognizer
9
9
ms.topic: conceptual
10
-
ms.date: 11/02/2021
10
+
ms.date: 03/11/2022
11
11
ms.author: lajanuar
12
12
recommendations: false
13
13
ms.custom: ignite-fall-2021
@@ -18,7 +18,7 @@ ms.custom: ignite-fall-2021
18
18
19
19
The business card model combines powerful Optical Character Recognition (OCR) capabilities with deep learning models to analyze and extract key information from business card images. The API analyzes printed business cards; extracts key information such as first name, last name, company name, email address, and phone number; and returns a structured JSON data representation.
20
20
21
-
***Sample business card processed with Form Recognizer Studio***
21
+
***Sample business card processed with [Form Recognizer Studio](https://formrecognizer.appliedai.azure.com/studio/prebuilt?formType=businessCard)***
22
22
23
23
:::image type="content" source="./media/studio/overview-business-card-studio.png" alt-text="sample business card" lightbox="./media/overview-business-card.jpg":::
24
24
@@ -64,7 +64,7 @@ See how data, including name, job title, address, email, and company name, is ex
64
64
65
65
#### Sample Labeling tool
66
66
67
-
You will need a business card document. You can use our [sample business card document](https://raw.githubusercontent.com/Azure-Samples/cognitive-services-REST-api-samples/master/curl/form-recognizer/businessCard.png).
67
+
You'll need a business card document. You can use our [sample business card document](https://raw.githubusercontent.com/Azure-Samples/cognitive-services-REST-api-samples/master/curl/form-recognizer/businessCard.png).
68
68
69
69
1. On the Sample Labeling tool home page, select **Use prebuilt model to get data**.
70
70
@@ -81,7 +81,7 @@ You will need a business card document. You can use our [sample business card do
81
81
* Supported file formats: JPEG, PNG, BMP, TIFF, and PDF (text-embedded or scanned). Text-embedded PDFs are best to eliminate the possibility of error in character extraction and location.
82
82
* For PDF and TIFF, up to 2000 pages can be processed (with a free tier subscription, only the first two pages are processed).
83
83
* The file size must be less than 50 MB.
84
-
* Image dimensions must be between 50 x 50 pixels and 10000 x 10000 pixels.
84
+
* Image dimensions must be between 50 x 50 pixels and 10,000 x 10,000 pixels.
85
85
* PDF dimensions are up to 17 x 17 inches, corresponding to Legal or A3 paper size, or smaller.
86
86
* The total size of the training data is 500 pages or less.
87
87
* If your PDFs are password-locked, you must remove the lock before submission.
Copy file name to clipboardExpand all lines: articles/applied-ai-services/form-recognizer/concept-id-document.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -7,7 +7,7 @@ manager: nitinme
7
7
ms.service: applied-ai-services
8
8
ms.subservice: forms-recognizer
9
9
ms.topic: conceptual
10
-
ms.date: 11/02/2021
10
+
ms.date: 03/11/2022
11
11
ms.author: lajanuar
12
12
recommendations: false
13
13
ms.custom: ignite-fall-2021
@@ -18,7 +18,7 @@ ms.custom: ignite-fall-2021
18
18
19
19
The ID document model combines Optical Character Recognition (OCR) with deep learning models to analyze and extracts key information from US Drivers Licenses (all 50 states and District of Columbia) and international passport biographical pages (excludes visa and other travel documents). The API analyzes identity documents, extracts key information, and returns a structured JSON data representation.
20
20
21
-
***Sample U.S. Driver's License processed with Form Recognizer Studio***
21
+
***Sample U.S. Driver's License processed with [Form Recognizer Studio](https://formrecognizer.appliedai.azure.com/studio/prebuilt?formType=idDocument)***
22
22
23
23
:::image type="content" source="media/studio/analyze-drivers-license.png" alt-text="Image of a sample driver's license." lightbox="media/overview-id.jpg":::
Copy file name to clipboardExpand all lines: articles/applied-ai-services/form-recognizer/concept-layout.md
+8-8Lines changed: 8 additions & 8 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -7,7 +7,7 @@ manager: nitinme
7
7
ms.service: applied-ai-services
8
8
ms.subservice: forms-recognizer
9
9
ms.topic: conceptual
10
-
ms.date: 11/02/2021
10
+
ms.date: 03/11/2022
11
11
ms.author: lajanuar
12
12
recommendations: false
13
13
ms.custom: ignite-fall-2021
@@ -17,9 +17,9 @@ ms.custom: ignite-fall-2021
17
17
18
18
The Form Recognizer Layout API extracts text, tables, selection marks, and structure information from documents (PDF, TIFF) and images (JPG, PNG, BMP).
19
19
20
-
***Sample form processed with [Form Recognizer Sample Labeling tool](https://fott-2-1.azurewebsites.net/)layout feature***
20
+
***Sample form processed with [Form Recognizer Studio](https://formrecognizer.appliedai.azure.com/studio/layout)***
:::image type="content" source="media/studio/analyze-layout.png" alt-text="Screenshot: Screenshot of sample document processed using Form Recognizer studio":::
23
23
24
24
**Data extraction features**
25
25
@@ -92,20 +92,20 @@ You'll need a form document. You can use our [sample form document](https://raw.
92
92
* Supported file formats: JPEG, PNG, BMP, TIFF, and PDF (text-embedded or scanned). Text-embedded PDFs are best to eliminate the possibility of error in character extraction and location.
93
93
* For PDF and TIFF, up to 2000 pages can be processed (with a free tier subscription, only the first two pages are processed).
94
94
* The file size must be less than 50 MB.
95
-
* Image dimensions must be between 50 x 50 pixels and 10000 x 10000 pixels.
95
+
* Image dimensions must be between 50 x 50 pixels and 10,000 x 10,000 pixels.
96
96
97
97
> [!NOTE]
98
98
> The [Sample Labeling tool](https://fott-2-1.azurewebsites.net/) does not support the BMP file format. This is a limitation of the tool not the Form Recognizer Service.
99
99
100
100
## Supported languages and locales
101
101
102
-
Form Recognizer preview version introduces additional language support for the layout model. *See*our [Language Support](language-support.md) for a complete list of supported handwritten and printed languages.
102
+
Form Recognizer preview version introduces additional language support for the layout model. *See*[Language Support](language-support.md) for a complete list of supported handwritten and printed languages.
103
103
104
104
## Features
105
105
106
106
### Tables and table headers
107
107
108
-
Layout API extracts tables in the `pageResults` section of the JSON output. Documents can be scanned, photographed, or digitized. Tables can be complex with merged cells or columns, with or without borders, and with odd angles. Extracted table information includes the number of columns and rows, row span, and column span. Each cell with its bounding box is output along with information whether it's recognized as part of a header or not. The model predicted header cells can span multiple rows and are not necessarily the first rows in a table. They also work with rotated tables. Each table cell also includes the full text with references to the individual words in the `readResults` section.
108
+
Layout API extracts tables in the `pageResults` section of the JSON output. Documents can be scanned, photographed, or digitized. Tables can be complex with merged cells or columns, with or without borders, and with odd angles. Extracted table information includes the number of columns and rows, row span, and column span. Each cell with its bounding box is output along with information whether it's recognized as part of a header or not. The model predicted header cells can span multiple rows and aren't necessarily the first rows in a table. They also work with rotated tables. Each table cell also includes the full text with references to the individual words in the `readResults` section.
@@ -117,15 +117,15 @@ Layout API also extracts selection marks from documents. Extracted selection mar
117
117
118
118
### Text lines and words
119
119
120
-
Layout API extracts text from documents and images with multiple text angles and colors. It accepts photos of documents, faxes, printed and/or handwritten (English only) text, and mixed modes. Text is extracted with information provided on lines, words, bounding boxes, confidence scores, and style (handwritten or other). All the text information is included in the `readResults` section of the JSON output.
120
+
Layout API extracts text from documents and images with multiple text angles and colors. It accepts photos of documents, faxes, printed and/or handwritten (English only) text, and mixed modes. Text is extracted with information provided in lines, words, and bounding boxes. All the text information is included in the `readResults` section of the JSON output.
121
121
122
122
:::image type="content" source="./media/layout-text-extraction.png" alt-text="Layout text extraction output":::
123
123
124
124
### Natural reading order for text lines (Latin only)
125
125
126
126
In Form Recognizer v2.1, you can specify the order in which the text lines are output with the `readingOrder` query parameter. Use `natural` for a more human-friendly reading order output as shown in the following example. This feature is only supported for Latin languages.
127
127
128
-
In Form Recognizer v3.0, the natural reading order output is used by the service in all cases. Therefore, there is no `readingOrder` parameter provided in this version.
128
+
In Form Recognizer v3.0, the natural reading order output is used by the service in all cases. Therefore, there's no `readingOrder` parameter provided in this version.
129
129
130
130
### Handwritten classification for text lines (Latin only)
Copy file name to clipboardExpand all lines: articles/applied-ai-services/form-recognizer/concept-receipt.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -7,7 +7,7 @@ manager: nitinme
7
7
ms.service: applied-ai-services
8
8
ms.subservice: forms-recognizer
9
9
ms.topic: conceptual
10
-
ms.date: 11/02/2021
10
+
ms.date: 03/11/2022
11
11
ms.author: lajanuar
12
12
recommendations: false
13
13
ms.custom: ignite-fall-2021
@@ -18,7 +18,7 @@ ms.custom: ignite-fall-2021
18
18
19
19
The receipt model combines powerful Optical Character Recognition (OCR) capabilities with deep learning models to analyze and extract key information from sales receipts. Receipts can be of various formats and quality including printed and handwritten receipts. The API extracts key information such as merchant name, merchant phone number, transaction date, tax, and transaction total and returns a structured JSON data representation.
20
20
21
-
***Sample receipt processed with [Form Recognizer Sample Labeling tool](https://fott-2-1.azurewebsites.net/)***:
21
+
***Sample receipt processed with [Form Recognizer Studio](https://formrecognizer.appliedai.azure.com/studio/prebuilt?formType=receipt)***:
0 commit comments