You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-services/document-intelligence/concept-analyze-document-response.md
+7-8Lines changed: 7 additions & 8 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -6,11 +6,10 @@ author: laujan
6
6
manager: nitinme
7
7
ms.service: azure-ai-document-intelligence
8
8
ms.topic: conceptual
9
-
ms.date: 08/07/2024
9
+
ms.date: 09/26/2024
10
10
ms.author: vikurpad
11
11
ms.custom:
12
12
- references_regions
13
-
- ignite-2023
14
13
monikerRange: '>=doc-intel-3.0.0'
15
14
---
16
15
@@ -22,7 +21,7 @@ In this article, let's examine the different objects returned as part of the `An
22
21
23
22
## Analyze document request
24
23
25
-
The Document Intelligence APIs analyze images, PDFs, and other document files to extract and detect various content, layout, style, and semantic elements. The analyze operation is an async API. Submitting a document returns an **Operation-Location** header that contains the URL to poll for completion. When an analysis request completes successfully, the response contains the elements described in the [model data extraction](concept-model-overview.md#model-data-extraction).
24
+
The Document Intelligence APIs analyze images, PDFs, and other document files to extract and detect various content, layout, style, and semantic elements. The `Analyze` operation is an async API. Submitting a document returns an **Operation-Location** header that contains the URL to poll for completion. When an analysis request completes successfully, the response contains the elements described in the [model data extraction](concept-model-overview.md#model-data-extraction).
26
25
27
26
### Response elements
28
27
@@ -43,7 +42,7 @@ The top-level content property contains a concatenation of all content elements
43
42
44
43
## Analyze response
45
44
46
-
The analyze response for each API returns different objects. API responses contain elements from component models where applicable.
45
+
The `Analyze` response for each API returns different objects. API responses contain elements from component models where applicable.
47
46
48
47
| Response content | Description | API |
49
48
|--|--|--|
@@ -52,8 +51,8 @@ The analyze response for each API returns different objects. API responses conta
52
51
|**styles**| Identified text element properties. | Read, Layout, General Document, Prebuilt, and Custom models|
53
52
|**languages**| Identified language associated with each span of the text extracted | Read |
54
53
|**tables**| Tabular content identified and extracted from the document. Tables relate to tables identified by the pretrained layout model. Content labeled as tables is extracted as structured fields in the documents object.| Layout, General Document, Invoice, and Custom models |
55
-
|**figures**| Figures (charts, images) identified and extracted from the document, providing visual representations that aid in the understanding of complex information. | Layout model |
56
-
|**sections**| Hierarchical document structure identified and extracted from the document. Section or subsection with the corresponding elements (paragraph, table, figure) attached to it. | Layout model |
54
+
|**figures**| Figures (charts, images) identified and extracted from the document, providing visual representations that aid in the understanding of complex information. |The Layout model |
55
+
|**sections**| Hierarchical document structure identified and extracted from the document. Section or subsection with the corresponding elements (paragraph, table, figure) attached to it. |The Layout model |
57
56
|**keyValuePairs**| Key-value pairs recognized by a pretrained model. The key is a span of text from the document with the associated value. | General document and Invoice models |
58
57
|**documents**| Fields recognized are returned in the `fields` dictionary within the list of documents| Prebuilt models, Custom models.|
59
58
@@ -142,7 +141,7 @@ Based on its position and styling, a cell can be classified as general content,
142
141
Figures (charts, images) in documents play a crucial role in complementing and enhancing the textual content, providing visual representations that aid in the understanding of complex information. The figures object detected by the Layout model has key properties like `boundingRegions` (the spatial locations of the figure on the document pages, including the page number and the polygon coordinates that outline the figure's boundary), `spans` (details the text spans related to the figure, specifying their offsets and lengths within the document's text. This connection helps in associating the figure with its relevant textual context), `elements` (the identifiers for text elements or paragraphs within the document that are related to or describe the figure) and `caption`, if any.
143
142
144
143
When *output=figures* is specified during the initial `Analyze` operation, the service generates cropped images for all detected figures that can be accessed via `/analyeResults/{resultId}/figures/{figureId}`.
145
-
`FigureId`will be included in each figure object, following an undocumented convention of `{pageNumber}.{figureIndex}` where `figureIndex` resets to one per page.
144
+
`FigureId`is included in each figure object, following an undocumented convention of `{pageNumber}.{figureIndex}` where `figureIndex` resets to one per page.
146
145
147
146
```json
148
147
{
@@ -255,7 +254,7 @@ The semantic schema of a document type is described via the fields it contains.
255
254
| date | Date | ISO 8601 - YYYY-MM-DD | InvoiceDate: "5/7/2022" → "2022-05-07" |
256
255
| time | Time | ISO 8601 - hh:mm:ss | TransactionTime: "9:45 PM" → "21:45:00" |
Copy file name to clipboardExpand all lines: articles/ai-services/document-intelligence/concept-custom-classifier.md
+5-6Lines changed: 5 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,16 +1,15 @@
1
1
---
2
-
title: Custom classification model - Document Intelligence
2
+
title: Custom classification model - Document Intelligence
3
3
titleSuffix: Azure AI services
4
4
description: Use the custom classification model to train a model to identify and split the documents you process within your application.
5
5
author: vkurpad
6
6
manager: nitinme
7
7
ms.service: azure-ai-document-intelligence
8
8
ms.topic: conceptual
9
-
ms.date: 08/07/2024
9
+
ms.date: 09/26/2024
10
10
ms.author: lajanuar
11
11
ms.custom:
12
12
- references_regions
13
-
- ignite-2023
14
13
monikerRange: '>=doc-intel-3.1.0'
15
14
---
16
15
@@ -50,7 +49,6 @@ Custom classification models are deep-learning-model types that combine layout a
50
49
Custom classification models can analyze a single- or multi-file documents to identify if any of the trained document types are contained within an input file. Here are the currently supported scenarios:
51
50
52
51
* A single file containing one document type, such as a loan application form.
53
-
54
52
* A single file containing multiple document types. For instance, a loan application package that contains a loan application form, payslip, and bank statement.
55
53
56
54
* A single file containing multiple instances of the same document. For instance, a collection of scanned invoices.
@@ -59,7 +57,8 @@ Custom classification models can analyze a single- or multi-file documents to id
59
57
60
58
✔️ The maximum allowed number of classes is `500`. The maximum allowed number of document samples per class is `100`.
61
59
62
-
The model classifies each page of the input document, unless specified, to one of the classes in the labeled dataset. You can specify the page numbers to analyze in the input document as well. To set the threshold for your application, use the confidence score from the response.
60
+
The model classifies each page of the input document, unless specified, to one of the classes in the labeled dataset. You can specify the page numbers to analyze in the input document as well. To set the threshold for your application, use the confidence score from the response.
61
+
63
62
### Incremental training
64
63
65
64
With custom models, you need to maintain access to the training dataset to update your classifier with new samples for an existing class, or add new classes. Classifier models now support incremental training where you can reference an existing classifier and append new samples for an existing class or add new classes with samples. Incremental training enables scenarios where data retention is a challenge and the classifier needs to be updated to align with changing business needs. Incremental training is supported with models trained with API version `2024-02-29-preview` and later.
@@ -250,7 +249,7 @@ Alternatively, if you have a flat list of files or only plan to use a few select
250
249
```
251
250
252
251
As an example, the file list `car-maint.jsonl` contains the following files.
253
-
252
+
254
253
```json
255
254
{"file":"classifier/car-maint/Commercial Motor Vehicle - Adatum.pdf"}
256
255
{"file":"classifier/car-maint/Commercial Motor Vehicle - Fincher.pdf"}
1. Next, you're going to fill out the **Create Document Intelligence** fields with the following values:
40
+
1. Next, you're going to fill out the **`Create Document Intelligence`** fields with the following values:
41
41
42
42
***Subscription**. Select your current subscription.
43
-
***Resource group**. The [Azure resource group](/azure/cloud-adoption-framework/govern/resource-consistency/resource-access-management#what-is-an-azure-resource-group) that contains your resource. You can create a new group or add it to a pre-existing group.
43
+
***Resource group**. The [Azure resource group](/azure/cloud-adoption-framework/govern/resource-consistency/resource-access-management#what-is-an-azure-resource-group) that contains your resource. You can create a new group or add it to an existing group.
44
44
***Region**. Select your local region.
45
45
***Name**. Enter a name for your resource. We recommend using a descriptive name, for example *YourNameFormRecognizer*.
46
46
***Pricing tier**. The cost of your resource depends on the pricing tier you choose and your usage. For more information, see [pricing details](https://azure.microsoft.com/pricing/details/cognitive-services/). You can use the free pricing tier (F0) to try the service, and upgrade later to a paid tier for production.
@@ -122,7 +122,7 @@ For detailed information regarding Azure AI Translator service request limits, *
122
122
123
123
1. Open the **Program.cs** file.
124
124
125
-
1. Delete the pre-existing code, including the line `Console.WriteLine("Hello World!")`. Copy and paste the code sample into your application's Program.cs file. Make sure you update the key variable with the value from your Azure portal Translator instance:
125
+
1. Delete the existing code, including the line `Console.WriteLine("Hello World!")`. Copy and paste the code sample into your application's Program.cs file. Make sure you update the key variable with the value from your Azure portal Translator instance:
126
126
127
127
```csharp
128
128
usingSystem.Text;
@@ -210,15 +210,15 @@ After a successful call, you should see the following response:
210
210
211
211
### Set up your Go environment
212
212
213
-
You can use any text editor to write Go applications. We recommend using the latest version of [Visual Studio Code and the Go extension](/azure/developer/go/configure-visual-studio-code).
213
+
You can use any text editor to write Go applications. We recommend using the latest version of [Visual Studio Code and Go extension](/azure/developer/go/configure-visual-studio-code).
214
214
215
215
> [!TIP]
216
216
>
217
217
> If you're new to Go, try the [Get started with Go](/training/modules/go-get-started/) Learn module.
218
218
219
219
1. Make sure the latest version of [Go](https://go.dev/doc/install) is installed:
220
220
221
-
* Download the Go version for your operating system.
221
+
* Download the `Go` programming language version for your operating system.
222
222
* Once the download is complete, run the installer.
223
223
* Open a command prompt and enter the following to confirm Go was installed:
224
224
@@ -362,7 +362,7 @@ After a successful call, you should see the following response:
362
362
363
363
*[**Gradle**](https://docs.gradle.org/current/userguide/installation.html), version 6.8 or later.
364
364
<!-- checked -->
365
-
<!--
365
+
<!--
366
366
> [!div class="nextstepaction"]
367
367
> [I ran into an issue](https://microsoft.qualtrics.com/jfe/form/SV_0Cl5zkG3CnDjq6O?PLanguage=Java&Product=Translator&Page=quickstart-translator&Section=set-up-your-java-environment) -->
368
368
@@ -633,7 +633,7 @@ Add the following code sample to your `index.js` file. **Make sure you update th
633
633
634
634
```
635
635
<!-- checked -->
636
-
<!--
636
+
<!--
637
637
> [!div class="nextstepaction"]
638
638
> [I ran into an issue](https://microsoft.qualtrics.com/jfe/form/SV_0Cl5zkG3CnDjq6O?PLanguage=Python&Product=Translator&Page=quickstart-translator&Section=build-your-javascript-application) -->
639
639
@@ -671,7 +671,7 @@ After a successful call, you should see the following response:
671
671
672
672
```
673
673
<!-- checked -->
674
-
<!--
674
+
<!--
675
675
> [!div class="nextstepaction"]
676
676
> [My REST API call was successful](#next-steps) [I ran into an issue](https://microsoft.qualtrics.com/jfe/form/SV_0Cl5zkG3CnDjq6O?PLanguage=Java&Product=Translator&Page=quickstart-translator&Section=run-your-javascript-application) -->
677
677
@@ -694,7 +694,7 @@ After a successful call, you should see the following response:
694
694
> [!NOTE]
695
695
> We will also use a Python built-in package called json. It's used to work with JSON data.
696
696
<!-- checked -->
697
-
<!--
697
+
<!--
698
698
> [!div class="nextstepaction"]
699
699
> [I ran into an issue](https://microsoft.qualtrics.com/jfe/form/SV_0Cl5zkG3CnDjq6O?PLanguage=Python&Product=Translator&Page=quickstart-translator&Section=set-up-your-python-project) -->
> [I ran into an issue](https://microsoft.qualtrics.com/jfe/form/SV_0Cl5zkG3CnDjq6O?PLanguage=Python&Product=Translator&Page=quickstart-translator&Section=build-your-python-application) -->
750
750
@@ -782,7 +782,7 @@ After a successful call, you should see the following response:
782
782
783
783
```
784
784
<!-- checked -->
785
-
<!--
785
+
<!--
786
786
> [!div class="nextstepaction"]
787
787
> [My REST API call was successful](#next-steps) [I ran into an issue](https://microsoft.qualtrics.com/jfe/form/SV_0Cl5zkG3CnDjq6O?PLanguage=Python&Product=Translator&Page=quickstart-translator&Section=run-your-python-application) -->
You can use the client directly or [#configure-the-models-used-by-your-code](Configure themodelsused by yourcode) in LlamaIndex. To use the model directly, use the `chat` method for chat instruction models:
141
+
You can use the client directly or [Configure themodelsused by yourcode](#configure-the-models-used-by-your-code) in LlamaIndex. To use the model directly, use the `chat` method for chat instruction models:
0 commit comments