Skip to content

Commit d348de5

Browse files
Merge pull request #230923 from laujan/remove-entity-reference
remove entity reference
2 parents 2854c46 + 81d01a9 commit d348de5

File tree

10 files changed

+26
-44
lines changed

10 files changed

+26
-44
lines changed

articles/applied-ai-services/form-recognizer/concept-analyze-document-response.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -64,7 +64,7 @@ Spans specify the logical position of each element in the overall reading order,
6464

6565
### Bounding Region
6666

67-
Bounding regions describe the visual position of each element in the file. Since elements may not be visually contiguous (entities) or may cross pages (tables), the positions of most elements are described via an array of bounding regions. Each region specifies the page number (`1`-indexed) and bounding polygon. The bounding polygon is described as a sequence of points, clockwise from the left relative to the natural orientation of the element. For quadrilaterals, plot points are top-left, top-right, bottom-right, and bottom-left corners. Each point is represented by its x, y coordinate in the page unit specified by the unit property. In general, unit of measure for images is pixels while PDFs use inches.
67+
Bounding regions describe the visual position of each element in the file. Since elements may not be visually contiguous or may cross pages (tables), the positions of most elements are described via an array of bounding regions. Each region specifies the page number (`1`-indexed) and bounding polygon. The bounding polygon is described as a sequence of points, clockwise from the left relative to the natural orientation of the element. For quadrilaterals, plot points are top-left, top-right, bottom-right, and bottom-left corners. Each point is represented by its x, y coordinate in the page unit specified by the unit property. In general, unit of measure for images is pixels while PDFs use inches.
6868

6969
:::image type="content" source="media/bounding-regions.png" alt-text="Screenshot of detected bounding regions example.":::
7070

articles/applied-ai-services/form-recognizer/concept-form-recognizer-studio.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@ The following Form Recognizer service features are available in the Studio.
3131

3232
* **Layout**: Try out Form Recognizer's Layout feature to extract text, tables, selection marks, and structure information. Start with the [Studio Layout feature](https://formrecognizer.appliedai.azure.com/studio/layout). Explore with sample documents and your documents. Use the interactive visualization and JSON output to understand how the feature works. See the [Layout overview](concept-layout.md) to learn more and get started with the [Python SDK quickstart for Layout](quickstarts/get-started-sdks-rest-api.md?view=form-recog-3.0.0&preserve-view=true#layout-model).
3333

34-
* **General Documents**: Try out Form Recognizer's General Documents feature to extract key-value pairs and entities. Start with the [Studio General Documents feature](https://formrecognizer.appliedai.azure.com/studio/document). Explore with sample documents and your documents. Use the interactive visualization and JSON output to understand how the feature works. See the [General Documents overview](concept-general-document.md) to learn more and get started with the [Python SDK quickstart for Layout](quickstarts/get-started-sdks-rest-api.md?view=form-recog-3.0.0&preserve-view=true#general-document-model).
34+
* **General Documents**: Try out Form Recognizer's General Documents feature to extract key-value pairs. Start with the [Studio General Documents feature](https://formrecognizer.appliedai.azure.com/studio/document). Explore with sample documents and your documents. Use the interactive visualization and JSON output to understand how the feature works. See the [General Documents overview](concept-general-document.md) to learn more and get started with the [Python SDK quickstart for Layout](quickstarts/get-started-sdks-rest-api.md?view=form-recog-3.0.0&preserve-view=true#general-document-model).
3535

3636
* **Prebuilt models**: Form Recognizer's prebuilt models enable you to add intelligent document processing to your apps and flows without having to train and build your own models. As an example, start with the [Studio Invoice feature](https://formrecognizer.appliedai.azure.com/studio/prebuilt?formType=invoice). Explore with sample documents and your documents. Use the interactive visualization, extracted fields list, and JSON output to understand how the feature works. See the [Models overview](concept-model-overview.md) to learn more and get started with the [Python SDK quickstart for Prebuilt Invoice](quickstarts/get-started-sdks-rest-api.md?view=form-recog-3.0.0&preserve-view=true#prebuilt-model).
3737

articles/applied-ai-services/form-recognizer/faq.yml

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -154,8 +154,6 @@ sections:
154154
155155
- For a PDF, coordinates are given in inches.
156156
157-
- You can use the bounding box returned by Form Recognizer to identify the location of recognized entities.
158-
159157
- name: App development
160158
questions:
161159
- question: |

articles/applied-ai-services/form-recognizer/quickstarts/includes/v3-csharp-sdk.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ recommendations: false
2020

2121
In this quickstart, you use the following features to analyze and extract data and values from forms and documents:
2222

23-
* [**General document model**](#general-document-model)—Analyze and extract text, tables, structure, key-value pairs, and named entities.
23+
* [**General document model**](#general-document-model)—Analyze and extract text, tables, structure, key-value pairs.
2424

2525
* [**Layout model**](#layout-model)—Analyze and extract tables, lines, words, and selection marks like radio buttons and check boxes in documents, without the need to train a model.
2626

@@ -121,7 +121,7 @@ dotnet run formrecognizer-quickstart.dll
121121

122122
## General document model
123123

124-
Analyze and extract text, tables, structure, key-value pairs, and named entities.
124+
Analyze and extract text, tables, structure, key-value pairs.
125125

126126
> [!div class="checklist"]
127127
>

articles/applied-ai-services/form-recognizer/quickstarts/includes/v3-java-sdk.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ recommendations: false
1717

1818
In this quickstart you'll, use the following features to analyze and extract data and values from forms and documents:
1919

20-
* [**General document**](#general-document-model)—Analyze and extract text, tables, structure, key-value pairs, and named entities.
20+
* [**General document**](#general-document-model)—Analyze and extract text, tables, structure, key-value pairs.
2121

2222
* [**Layout**](#layout-model)—Analyze and extract tables, lines, words, and selection marks like radio buttons and check boxes in documents, without the need to train a model.
2323

@@ -138,7 +138,7 @@ To interact with the Form Recognizer service, you need to create an instance of
138138

139139
## General document model
140140

141-
Extract text, tables, structure, key-value pairs, and named entities from documents.
141+
Extract text, tables, structure, and key-value pairs from documents.
142142

143143
> [!div class="checklist"]
144144
>

articles/applied-ai-services/form-recognizer/quickstarts/includes/v3-javascript-sdk.md

Lines changed: 2 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ recommendations: false
1717

1818
In this quickstart you'll, use the following features to analyze and extract data and values from forms and documents:
1919

20-
* [**General document**](#general-document-model)—Analyze and extract key-value pairs, selection marks, and entities from documents.
20+
* [**General document**](#general-document-model)—Analyze and extract key-value pairs, and selection marks from documents.
2121

2222
* [**Layout**](#layout-model)—Analyze and extract tables, lines, words, and selection marks like radio buttons and check boxes in documents, without the need to train a model.
2323

@@ -102,7 +102,7 @@ To interact with the Form Recognizer service, you need to create an instance of
102102

103103
## General document model
104104

105-
Extract text, tables, structure, key-value pairs, and named entities from documents.
105+
Extract text, tables, structure,and key-value pairs from documents.
106106

107107
> [!div class="checklist"]
108108
>
@@ -184,12 +184,6 @@ Key-Value Pairs:
184184
Value: "Common Stock, $0.00000625 par value per share" (0.748)
185185
- Key : "Outstanding as of April 24, 2020"
186186
Value: "7,583,440,247 shares" (0.838)
187-
Entities:
188-
- "$0.00000625" Quantity - Currency (0.8)
189-
- "MSFT" Organization - <none> (0.99)
190-
- "NASDAQ" Organization - StockExchange (0.99)
191-
- "2.125%" Quantity - Percentage (0.8)
192-
- "2021" DateTime - DateRange (0.8)
193187
```
194188

195189
To view the entire output, visit the Azure samples repository on GitHub to view the [general document model output](https://github.com/Azure-Samples/cognitive-services-quickstart-code/blob/master/javascript/FormRecognizer/v3-javascript-sdk-general-document-output.md)

articles/applied-ai-services/form-recognizer/quickstarts/includes/v3-python-sdk.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ recommendations: false
1717

1818
In this quickstart you'll, use the following features to analyze and extract data and values from forms and documents:
1919

20-
* [**General document**](#general-document-model)—Analyze and extract text, tables, structure, key-value pairs, and named entities.
20+
* [**General document**](#general-document-model)—Analyze and extract text, tables, structure, and key-value pairs.
2121

2222
* [**Layout**](#layout-model)—Analyze and extract tables, lines, words, and selection marks like radio buttons and check boxes in documents, without the need to train a model.
2323

@@ -77,7 +77,7 @@ To interact with the Form Recognizer service, you need to create an instance of
7777

7878
## General document model
7979

80-
Extract text, tables, structure, key-value pairs, and named entities from documents.
80+
Extract text, tables, structure, and key-value pairs from documents.
8181

8282
> [!div class="checklist"]
8383
>

articles/applied-ai-services/form-recognizer/quickstarts/try-form-recognizer-studio.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ Prebuilt models help you add Form Recognizer features to your apps without havin
3636

3737
#### Document analysis
3838

39-
* [**General document**](https://formrecognizer.appliedai.azure.com/studio/prebuilt?formType=document): extract text, tables, structure, key-value pairs and named entities.
39+
* [**General document**](https://formrecognizer.appliedai.azure.com/studio/prebuilt?formType=document): extract text, tables, structure, key-value pairs.
4040
* [**Layout**](https://formrecognizer.appliedai.azure.com/studio/layout): extract text, tables, selection marks, and structure information from documents (PDF, TIFF) and images (JPG, PNG, BMP).
4141
* [**Read**](https://formrecognizer.appliedai.azure.com/studio/read): extract text lines, words, their locations, detected languages, and handwritten style if detected from documents (PDF, TIFF) and images (JPG, PNG, BMP).
4242

articles/applied-ai-services/form-recognizer/v3-migration-guide.md

Lines changed: 2 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@ The `2022-08-31` API has a few updates from the preview API versions:
2929
* Field deleted: entities removed from the result of the general document model.
3030
* Field rename: documentLanguage.languageCode to locale
3131
* Added support for HEIF format
32-
* Added paragraph detection, with role classification for layout and general document models
32+
* Added paragraph detection, with role classification for layout and general document models.
3333
* Added support for parsed address fields.
3434

3535
## Migrating from v2.1
@@ -224,17 +224,7 @@ Analyze response has been refactored to the following top-level results to suppo
224224
},
225225
"confidence": 0.95 // Extraction confidence
226226
}, ...
227-
], // List of extracted entities
228-
"entities": [
229-
{
230-
"category": "DateTime", // Primary entity category
231-
"subCategory": "Date", // Secondary entity category
232-
"content": "11/15/2019", // Entity content
233-
"boundingRegions": [ ... ], // Entity bounding regions
234-
"spans": [ ... ], // Entity spans
235-
"confidence": 0.99 // Extraction confidence
236-
}, ...
237-
], // List of extracted styles
227+
],
238228
"styles": [
239229
{
240230
"isHandwritten": true, // Is content in this style handwritten?

articles/applied-ai-services/form-recognizer/whats-new.md

Lines changed: 13 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -360,7 +360,7 @@ Form Recognizer service is updated on an ongoing basis. Bookmark this page to st
360360

361361
* **AI quality improvements**
362362

363-
* [**prebuilt-read**](concept-read.md). Enhanced support for single characters, handwritten dates, amounts, names, other entities commonly found in receipts and invoices and improved processing of digital PDF documents.
363+
* [**prebuilt-read**](concept-read.md). Enhanced support for single characters, handwritten dates, amounts, names, other key data commonly found in receipts and invoices and improved processing of digital PDF documents.
364364
* [**prebuilt-layout**](concept-layout.md). Support for better detection of cropped tables, borderless tables, and improved recognition of long spanning cells.
365365
* [**prebuilt-document**](concept-general-document.md). Improved value and check box detection.
366366
* [**custom-neural**](concept-custom-neural.md). Improved accuracy for table detection and extraction.
@@ -477,10 +477,10 @@ Form Recognizer service is updated on an ongoing basis. Bookmark this page to st
477477

478478
* Form Recognizer v3.0 preview release introduces several new features, capabilities and enhancements:
479479

480-
* [**Custom neural model**](concept-custom-neural.md) or custom document model is a new custom model to extract text and selection marks from structured forms, semi-strutured and **unstructured documents**.
480+
* [**Custom neural model**](concept-custom-neural.md) or custom document model is a new custom model to extract text and selection marks from structured forms, semi-structured and **unstructured documents**.
481481
* [**W-2 prebuilt model**](concept-w2.md) is a new prebuilt model to extract fields from W-2 forms for tax reporting and income verification scenarios.
482482
* [**Read**](concept-read.md) API extracts printed text lines, words, text locations, detected languages, and handwritten text, if detected.
483-
* [**General document**](concept-general-document.md) pretrained model is now updated to support selection marks in addition to API text, tables, structure, key-value pairs, and named entities from forms and documents.
483+
* [**General document**](concept-general-document.md) pretrained model is now updated to support selection marks in addition to API text, tables, structure, and key-value pairs from forms and documents.
484484
* [**Invoice API**](language-support.md#invoice-model) Invoice prebuilt model expands support to Spanish invoices.
485485
* [**Form Recognizer Studio**](https://formrecognizer.appliedai.azure.com) adds new demos for Read, W2, Hotel receipt samples, and support for training the new custom neural models.
486486
* [**Language Expansion**](language-support.md) Form Recognizer Read, Layout, and Custom Form add support for 42 new languages including Arabic, Hindi, and other languages using Arabic and Devanagari scripts to expand the coverage to 164 languages. Handwritten language support expands to Japanese and Korean.
@@ -544,7 +544,7 @@ Form Recognizer service is updated on an ongoing basis. Bookmark this page to st
544544

545545
* **Form Recognizer v3.0 preview release version 4.0.0-beta.1 (2021-10-07)introduces several new features and capabilities:**
546546

547-
* [**General document**](concept-general-document.md) model is a new API that uses a pretrained model to extract text, tables, structure, key-value pairs, and named entities from forms and documents.
547+
* [**General document**](concept-general-document.md) model is a new API that uses a pretrained model to extract text, tables, structure, and key-value pairs from forms and documents.
548548
* [**Hotel receipt**](concept-receipt.md) model added to prebuilt receipt processing.
549549
* [**Expanded fields for ID document**](concept-id-document.md) the ID model supports endorsements, restrictions, and vehicle classification extraction from US driver's licenses.
550550
* [**Signature field**](concept-custom.md) is a new field type in custom forms to detect the presence of a signature in a form field.
@@ -556,15 +556,15 @@ Form Recognizer service is updated on an ongoing basis. Bookmark this page to st
556556

557557
* Form Recognizer model data extraction
558558

559-
| **Model** | **Text extraction** |**Key-Value pairs** |**Selection Marks** | **Tables** |**Entities** |
560-
| --- | :---: |:---:| :---: | :---: |:---: |
561-
|General document ||||||
562-
| Layout || ||| |
563-
| Invoice ||||||
564-
|Receipt ||| | ||
565-
| ID document ||| | ||
566-
| Business card ||| | ||
567-
| Custom ||||||
559+
| **Model** | **Text extraction** |**Key-Value pairs** |**Selection Marks** | **Tables** |
560+
| --- | :---: |:---:| :---: | :---: |
561+
|General document |||||
562+
| Layout || |||
563+
| Invoice |||||
564+
|Receipt ||| | |
565+
| ID document ||| | |
566+
| Business card ||| | |
567+
| Custom |||||
568568

569569
---
570570

0 commit comments

Comments
 (0)