You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/applied-ai-services/form-recognizer/concept-custom.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -123,7 +123,7 @@ The [Sample Labeling tool](https://fott-2-1.azurewebsites.net/) doesn't support
123
123
124
124
## Supported languages and locales
125
125
126
-
The Form Recognizer preview version introduces more language support for custom models. For a list of supported handwritten and printed text, see [Language support](language-support.md#layout-and-custom-model).
126
+
The Form Recognizer preview version introduces more language support for custom models. For a list of supported handwritten and printed text, see [Language support](language-support.md).
Copy file name to clipboardExpand all lines: articles/applied-ai-services/form-recognizer/concept-form-recognizer-studio.md
+6-4Lines changed: 6 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -27,13 +27,15 @@ The following image shows the Invoice prebuilt model feature at work.
27
27
28
28
The following Form Recognizer service features are available in the Studio.
29
29
30
-
***Layout**: Try out Form Recognizer's Layout feature to extract text, tables, selection marks, and structure information from documentsβPDF, TIFFβand imagesβJPG, PNG, BMP. Start with the [Studio Layout quickstart](quickstarts/try-v3-form-recognizer-studio.md#layout). Explore with sample documents and your documents. Use the interactive visualization and JSON output to understand how the feature works. See the [Layout overview](concept-layout.md) to learn more and get started with the [Python SDK quickstart for Layout](quickstarts/try-v3-python-sdk.md#layout-model).
30
+
***Read**: Try out Form Recognizer's Read feature to extract text lines, words, detected languages, and handwritten style if detected. Start with the [Studio Read feature](https://formrecognizer.appliedai.azure.com/studio/read). Explore with sample documents and your documents. Use the interactive visualization and JSON output to understand how the feature works. See the [Read overview](concept-read.md) to learn more and get started with the [Python SDK quickstart for Layout](quickstarts/try-v3-python-sdk.md).
31
31
32
-
***Prebuilt models**: Form Recognizer's pre-built models enable you to add intelligent document processing to your apps and flows without having to train and build your own models. Start with the [Studio Prebuilts quickstart](quickstarts/try-v3-form-recognizer-studio.md#prebuilt-models). Explore with sample documents and your documents. Use the interactive visualization, extracted fields list, and JSON output to understand how the feature works. See the [Models overview](concept-model-overview.md) to learn more and get started with the [Python SDK quickstart for Prebuilt Invoice](quickstarts/try-v3-python-sdk.md#prebuilt-model).
32
+
***Layout**: Try out Form Recognizer's Layout feature to extract text, tables, selection marks, and structure information. Start with the [Studio Layout feature](https://formrecognizer.appliedai.azure.com/studio/layout). Explore with sample documents and your documents. Use the interactive visualizationand JSON output to understand how the feature works. See the [Layout overview](concept-layout.md) to learn more and get started with the [Python SDK quickstart for Layout](quickstarts/try-v3-python-sdk.md#layout-model).
33
33
34
-
***Custom models**: Form Recognizer's custom models enable you to extract fields and values from models trained with your data, tailored to your forms and documents. Create standalone custom models or combine two or more custom models to create a composed model to extract data from multiple form types. Start with the [Studio Custom models quickstart](quickstarts/try-v3-form-recognizer-studio.md#custom-projects). Use the online wizard, labeling interface, training step, and visualizations to understand how the feature works. Test the custom model with your sample documents and iterate to improve the model. See the [Custom models overview](concept-custom.md) to learn more and use the [Form Recognizer v3.0 preview migration guide](v3-migration-guide.md) to start integrating the new models with your applications.
34
+
***General Documents**: Try out Form Recognizer's General Documents feature to extract key-value pairs and entities. Start with the [Studio General Documents feature](https://formrecognizer.appliedai.azure.com/studio/document). Explore with sample documents and your documents. Use the interactive visualization and JSON output to understand how the feature works. See the [General Documents overview](concept-general-document.md) to learn more and get started with the [Python SDK quickstart for Layout](quickstarts/try-v3-python-sdk.md#general-document-model).
35
35
36
-
***Custom models: Labeling features**: Form Recognizer Custom model creation requires identifying the fields to be extracted and labeling those fields before training the custom models. Labeling text, selection marks, tabular data, and other content types are typically assisted with a user interface to ease the training workflow. For example, use the [Label as tables](quickstarts/try-v3-form-recognizer-studio.md#labeling-as-tables) and [Labeling for signature detection](quickstarts/try-v3-form-recognizer-studio.md#labeling-for-signature-detection) quickstarts to understand the labeling experience in Form Recognizer Studio.
36
+
***Prebuilt models**: Form Recognizer's pre-built models enable you to add intelligent document processing to your apps and flows without having to train and build your own models. As an example, start with the [Studio Invoice feature](https://formrecognizer.appliedai.azure.com/studio/prebuilt?formType=invoice). Explore with sample documents and your documents. Use the interactive visualization, extracted fields list, and JSON output to understand how the feature works. See the [Models overview](concept-model-overview.md) to learn more and get started with the [Python SDK quickstart for Prebuilt Invoice](quickstarts/try-v3-python-sdk.md#prebuilt-model).
37
+
38
+
***Custom models**: Form Recognizer's custom models enable you to extract fields and values from models trained with your data, tailored to your forms and documents. Create standalone custom models or combine two or more custom models to create a composed model to extract data from multiple form types. Start with the [Studio Custom models feature](https://formrecognizer.appliedai.azure.com/studio/custommodel/projects). Use the online wizard, labeling interface, training step, and visualizations to understand how the feature works. Test the custom model with your sample documents and iterate to improve the model. See the [Custom models overview](concept-custom.md) to learn more and use the [Form Recognizer v3.0 preview migration guide](v3-migration-guide.md) to start integrating the new models with your applications.
Copy file name to clipboardExpand all lines: articles/applied-ai-services/form-recognizer/concept-layout.md
+7-17Lines changed: 7 additions & 17 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -15,7 +15,7 @@ ms.custom: ignite-fall-2021
15
15
16
16
# Form Recognizer layout model
17
17
18
-
Azure the Form Recognizer Layout API extracts text, tables, selection marks, and structure information from documents (PDF, TIFF) and images (JPG, PNG, BMP). The layout model combines an enhanced version of our powerful [Optical Character Recognition (OCR)](../../cognitive-services/computer-vision/overview-ocr.md) capabilities with deep learning models to extract text, tables, selection marks, and document structure.
18
+
The Form Recognizer Layout API extracts text, tables, selection marks, and structure information from documents (PDF, TIFF) and images (JPG, PNG, BMP).
19
19
20
20
***Sample form processed with [Form Recognizer Sample Labeling tool](https://fott-2-1.azurewebsites.net/) layout feature***
21
21
@@ -58,7 +58,7 @@ See how data, including tables, check boxes, and text, is extracted from forms a
58
58
59
59
***Sample form processed with [Form Recognizer Studio](https://formrecognizer.appliedai.azure.com/studio/layout)***
60
60
61
-
:::image type="content" source="media/studio/sample-layout.png" alt-text="Screenshot: document processing in Form Recognizer Studio.":::
61
+
:::image type="content" source="media/studio/form-recognizer-studio-layout-v3p2.png" alt-text="Screenshot: Layout processing in Form Recognizer Studio.":::
62
62
63
63
1. On the Form Recognizer Studio home page, select **Layout**
64
64
@@ -93,19 +93,13 @@ You'll need a form document. You can use our [sample form document](https://raw.
93
93
* For PDF and TIFF, up to 2000 pages can be processed (with a free tier subscription, only the first two pages are processed).
94
94
* The file size must be less than 50 MB.
95
95
* Image dimensions must be between 50 x 50 pixels and 10000 x 10000 pixels.
96
-
* PDF dimensions are up to 17 x 17 inches, corresponding to Legal or A3 paper size, or smaller.
97
-
* The total size of the training data is 500 pages or less.
98
-
* If your PDFs are password-locked, you must remove the lock before submission.
99
-
* For unsupervised learning (without labeled data):
100
-
* Data must contain keys and values.
101
-
* Keys must appear above or to the left of the values; they can't appear below or to the right.
102
96
103
97
> [!NOTE]
104
98
> The [Sample Labeling tool](https://fott-2-1.azurewebsites.net/) does not support the BMP file format. This is a limitation of the tool not the Form Recognizer Service.
105
99
106
100
## Supported languages and locales
107
101
108
-
Form Recognizer preview version introduces additional language support for the layout model. *See* our [Language Support](language-support.md#layout-and-custom-model) for a complete list of supported handwritten and printed text.
102
+
Form Recognizer preview version introduces additional language support for the layout model. *See* our [Language Support](language-support.md) for a complete list of supported handwritten and printed languages.
109
103
110
104
## Features
111
105
@@ -129,21 +123,17 @@ Layout API extracts text from documents and images with multiple text angles and
129
123
130
124
### Natural reading order for text lines (Latin only)
131
125
132
-
You can specify the order in which the text lines are output with the `readingOrder` query parameter. Use `natural` for a more human-friendly reading order output as shown in the following example. This feature is only supported for Latin languages.
126
+
In Form Recognizer v2.1, you can specify the order in which the text lines are output with the `readingOrder` query parameter. Use `natural` for a more human-friendly reading order output as shown in the following example. This feature is only supported for Latin languages.
133
127
134
-
:::image type="content" source="./media/layout-reading-order-example.png" alt-text="Layout Reading order example" lightbox="../../cognitive-services/Computer-vision/Images/ocr-reading-order-example.png":::
128
+
In Form Recognizer v3.0, the natural readingorder output is used by the service in all cases. Therefore, there is no `readingOrder` parameter provided in this version.
135
129
136
130
### Handwritten classification for text lines (Latin only)
137
131
138
-
The response includes classifying whether each text line is of handwriting style or not, along with a confidence score. This feature is only supported for Latin languages. The following example shows the handwritten classification for the text in the image.
The response includes classifying whether each text line is of handwriting style or not, along with a confidence score. This feature is only supported for Latin languages.
141
133
142
134
### Select page numbers or ranges for text extraction
143
135
144
-
For large multi-page documents, use the `pages` query parameter to indicate specific page numbers or page ranges for text extraction. The following example shows a document with 10 pages, with text extracted for both cases - all pages (1-10) and selected pages (3-6).
Copy file name to clipboardExpand all lines: articles/applied-ai-services/form-recognizer/concept-model-overview.md
+20-3Lines changed: 20 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -16,12 +16,13 @@ ms.custom: ignite-fall-2021
16
16
17
17
# Form Recognizer models
18
18
19
-
Azure Form Recognizer prebuilt models enable you to add intelligent document processing to your apps and flows without having to train and build your own models. Prebuilt models use optical character recognition (OCR) combined with deep learning models to identify and extract predefined text and data fields common to specific form and document types. Form Recognizer extracts analyzes form and document data then returns an organized, structured JSON response. Form Recognizer v2.1 supports invoice, receipt, ID document, and business card models.
19
+
Azure Form Recognizer prebuilt models enable you to add intelligent document processing to your apps and flows without having to train and build your own models. Prebuilt models use optical character recognition (OCR) combined with deep learning models to identify and extract predefined text and data fields common to specific form and document types. Form Recognizer extracts analyzes form and document data then returns an organized, structured JSON response. Form Recognizer v2.1 supports invoice, receipt, ID document, and business card models.
20
20
21
21
## Model overview
22
22
23
23
|**Model**|**Description**|
24
24
| --- | --- |
25
+
| π[Read (preview)](#read-preview)| Extract text lines, words, their locations, detected languages, and handwritten style if detected. |
25
26
| π[General document (preview)](#general-document-preview)| Extract text, tables, structure, key-value pairs, and named entities. |
26
27
|[Layout](#layout)| Extracts text and layout information from documents. |
27
28
|[Invoice](#invoice)| Extract key information from English and Spanish invoices. |
@@ -30,6 +31,19 @@ ms.custom: ignite-fall-2021
30
31
|[Business card](#business-card)| Extract key information from English business cards. |
31
32
|[Custom](#custom)| Extract data from forms and documents specific to your business. Custom models are trained for your distinct data and use cases. |
32
33
34
+
### Read (preview)
35
+
36
+
:::image type="content" source="media/studio/read-card.png" alt-text="Screenshot: Studio read icon.":::
37
+
38
+
The Read API analyzes and extracts ext lines, words, their locations, detected languages, and handwritten style if detected.
39
+
40
+
***Sample document processed using the [Form Recognizer Studio](https://formrecognizer.appliedai.azure.com/studio/read)***:
41
+
42
+
:::image type="content" source="media/studio/form-recognizer-studio-read-v3p2.png" alt-text="Screenshot: Screenshot of sample document processed using Form Recognizer studio Read":::
43
+
44
+
> [!div class="nextstepaction"]
45
+
> [Learn more: read model](concept-read.md)
46
+
33
47
### General document (preview)
34
48
35
49
:::image type="content" source="media/studio/general-document.png" alt-text="Screenshot: Studio general document icon.":::
@@ -128,15 +142,17 @@ The custom model analyzes and extracts data from forms and documents specific to
* For best results, provide one clear photo or high-quality scan per document.
@@ -158,6 +174,7 @@ The custom model analyzes and extracts data from forms and documents specific to
158
174
159
175
Form Recognizer v3.0 (preview) introduces several new features and capabilities:
160
176
177
+
*[**Read (preview)**](concept-read.md) model is a new API that extracts text lines, words, their locations, detected languages, and handwrting style if detected.
161
178
*[**General document (preview)**](concept-general-document.md) model is a new API that uses a pre-trained model to extract text, tables, structure, key-value pairs, and named entities from forms and documents.
162
179
*[**Receipt (preview)**](concept-receipt.md) model supports single-page hotel receipt processing.
163
180
*[**ID document (preview)**](concept-id-document.md) model supports endorsements, restrictions, and vehicle classification extraction from US driver's licenses.
0 commit comments