You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/applied-ai-services/form-recognizer/choose-model-feature.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -59,4 +59,4 @@ The following decision charts highlight the features of each **Form Recognizer v
59
59
60
60
## Next steps
61
61
62
-
*[Learn how to process your own forms and documents](quickstarts/try-v3-form-recognizer-studio.md) with the [Form Recognizer Studio](https://formrecognizer.appliedai.azure.com/studio)
62
+
*[Learn how to process your own forms and documents](quickstarts/try-form-recognizer-studio.md) with the [Form Recognizer Studio](https://formrecognizer.appliedai.azure.com/studio)
[Form Recognizer Studio](https://formrecognizer.appliedai.azure.com/) is an online tool for visually exploring, understanding, and integrating features from the Form Recognizer service into your applications. Use the [Form Recognizer Studio quickstart](quickstarts/try-v3-form-recognizer-studio.md) to get started analyzing documents with pretrained models. Build custom template models and reference the models in your applications using the [Python SDK v3.0](quickstarts/get-started-sdks-rest-api.md?view=form-recog-3.0.0&preserve-view=true) and other quickstarts.
19
+
[Form Recognizer Studio](https://formrecognizer.appliedai.azure.com/) is an online tool for visually exploring, understanding, and integrating features from the Form Recognizer service into your applications. Use the [Form Recognizer Studio quickstart](quickstarts/try-form-recognizer-studio.md) to get started analyzing documents with pretrained models. Build custom template models and reference the models in your applications using the [Python SDK v3.0](quickstarts/get-started-sdks-rest-api.md?view=form-recog-3.0.0&preserve-view=true) and other quickstarts.
20
20
21
21
The following image shows the Invoice prebuilt model feature at work.
22
22
@@ -43,4 +43,4 @@ The following Form Recognizer service features are available in the Studio.
43
43
* Refer to our [**v3.0 REST API quickstarts**](quickstarts/get-started-sdks-rest-api.md?view=form-recog-3.0.0&preserve-view=true) to try the v3.0features using the new REST API.
44
44
45
45
> [!div class="nextstepaction"]
46
-
> [Form Recognizer Studio quickstart](quickstarts/try-v3-form-recognizer-studio.md)
46
+
> [Form Recognizer Studio quickstart](quickstarts/try-form-recognizer-studio.md)
Copy file name to clipboardExpand all lines: articles/applied-ai-services/form-recognizer/concept-layout.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -7,7 +7,7 @@ manager: nitinme
7
7
ms.service: applied-ai-services
8
8
ms.subservice: forms-recognizer
9
9
ms.topic: conceptual
10
-
ms.date: 05/23/2023
10
+
ms.date: 06/23/2023
11
11
ms.author: lajanuar
12
12
---
13
13
@@ -447,7 +447,7 @@ See here for a [sample document file](https://github.com/Azure-Samples/cognitive
447
447
The JSON output has two parts:
448
448
449
449
*`readResults` node contains all of the recognized text and selection mark. The text presentation hierarchy is page, then line, then individual words.
450
-
*`pageResults` node contains the tables and cells extracted with their bounding boxes, confidence, and a reference to the lines and words in "readResults".
450
+
*`pageResults` node contains the tables and cells extracted with their bounding boxes, confidence, and a reference to the lines and words in "readResults" field.
451
451
452
452
## Example Output
453
453
@@ -474,7 +474,7 @@ Layout API also extracts selection marks from documents. Extracted selection mar
474
474
475
475
::: moniker range="form-recog-3.0.0"
476
476
477
-
*[Learn how to process your own forms and documents](quickstarts/try-v3-form-recognizer-studio.md) with the [Form Recognizer Studio](https://formrecognizer.appliedai.azure.com/studio)
477
+
*[Learn how to process your own forms and documents](quickstarts/try-form-recognizer-studio.md) with the [Form Recognizer Studio](https://formrecognizer.appliedai.azure.com/studio)
478
478
479
479
* Complete a [Form Recognizer quickstart](quickstarts/get-started-sdks-rest-api.md?view=form-recog-3.0.0&preserve-view=true) and get started creating a document processing app in the development language of your choice.
Copy file name to clipboardExpand all lines: articles/applied-ai-services/form-recognizer/faq.yml
+7-7Lines changed: 7 additions & 7 deletions
Original file line number
Diff line number
Diff line change
@@ -7,7 +7,7 @@ metadata:
7
7
ms.service: applied-ai-services
8
8
ms.subservice: forms-recognizer
9
9
ms.topic: faq
10
-
ms.date: 03/03/2023
10
+
ms.date: 06/23/2023
11
11
ms.author: lajanuar
12
12
monikerRange: '>=form-recog-2.1.0'
13
13
@@ -85,7 +85,7 @@ sections:
85
85
86
86
Models have the same lifecycle as the API version used to train the model. Custom models trained with a generally available (GA) version of the API have the same lifecycle as the API version. When the API version is deprecated, the model is no longer available for inference. Models trained with a preview version of the API also have the same lifecycle as the preview API.
87
87
88
-
Expect that preview APIs will be deprecated within three months of an updated preview API version or newer generally available API version.
88
+
Expect preview API deprecation within three months of an updated preview API version or newer generally available API version.
89
89
90
90
- question: |
91
91
What is the accuracy score and how is it calculated?
For more information, *see* [**Supported clients**](sdk-overview.md#supported-clients)
189
+
For more information, *see* [Supported clients](sdk-overview.md#supported-clients)
190
190
191
191
192
192
- question: |
@@ -371,7 +371,7 @@ sections:
371
371
- question: |
372
372
I'm building a custom model, what does the signature-detection label return?
373
373
answer: |
374
-
[Signature detection](quickstarts/try-v3-form-recognizer-studio.md#signature-detection) looks for the presence of a signature, not the identity of the person signing the document.
374
+
[Signature detection](quickstarts/try-form-recognizer-studio.md#signature-detection) looks for the presence of a signature, not the identity of the person signing the document.
375
375
376
376
If the model returns "unsigned" for signature detection, the model didn’t find a signature in the defined field.
377
377
@@ -386,7 +386,7 @@ sections:
386
386
387
387
- Do your tables span across multiple pages? If so, to avoid having to label all of the pages, split the PDF into pages prior to sending it to Form Recognizer. Following the analysis, post-process the pages to a single table.
388
388
389
-
- If you’re creating custom models, refer to [Labeling as tables](quickstarts/try-v3-form-recognizer-studio.md#labeling-as-tables). Dynamic tables have a variable number of rows for each given column. Fixed tables have a constant number of rows for each given column.
389
+
- If you’re creating custom models, refer to [Labeling as tables](quickstarts/try-form-recognizer-studio.md#labeling-as-tables). Dynamic tables have a variable number of rows for each given column. Fixed tables have a constant number of rows for each given column.
390
390
391
391
- question: |
392
392
How can I move my trained models from one environment (like beta) to another (like production)?
@@ -464,7 +464,7 @@ sections:
464
464
Form Recognizer connected containers send billing information to Azure by using a Form Recognizer resource on your Azure account. Connected containers don't send customer data, such as the image or text that's being analyzed, to Microsoft. See the [Cognitive Services container FAQ](../../cognitive-services/containers/disconnected-container-faq.yml#how-does-billing-work) for an example of the information sent to Microsoft for billing.
465
465
466
466
- question: |
467
-
I received an "OutOfQuota" error message: "Container isn't in a valid state. Subscription validation failed with status 'OutOfQuota'. API key is out of quota".
467
+
I received an "OutOfQuota" error message: *Container isn't in a valid state. Subscription validation failed with status 'OutOfQuota' API key is out of quota*.
468
468
answer: |
469
469
470
470
Form Recognizer connected containers send billing information to Azure by using a Form Recognizer resource on your Azure account. You could get this message if the containers can't communicate with the billing endpoint.
Copy file name to clipboardExpand all lines: articles/applied-ai-services/form-recognizer/how-to-guides/build-a-custom-classifier.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -45,7 +45,7 @@ Once you've put together the set of forms or documents for training, you need to
45
45
46
46
The Form Recognizer Studio provides and orchestrates all the API calls required to complete your dataset and train your model.
47
47
48
-
1. Start by navigating to the [Form Recognizer Studio](https://formrecognizer.appliedai.azure.com/studio). The first time you use the Studio, you need to [initialize your subscription, resource group, and resource](../quickstarts/try-v3-form-recognizer-studio.md). Then, follow the [prerequisites for custom projects](../quickstarts/try-v3-form-recognizer-studio.md#additional-prerequisites-for-custom-projects) to configure the Studio to access your training dataset.
48
+
1. Start by navigating to the [Form Recognizer Studio](https://formrecognizer.appliedai.azure.com/studio). The first time you use the Studio, you need to [initialize your subscription, resource group, and resource](../quickstarts/try-form-recognizer-studio.md). Then, follow the [prerequisites for custom projects](../quickstarts/try-form-recognizer-studio.md#added-prerequisites-for-custom-projects) to configure the Studio to access your training dataset.
49
49
50
50
1. In the Studio, select the **Custom classification model** tile, on the custom models section of the page and select the **Create a project** button.
Copy file name to clipboardExpand all lines: articles/applied-ai-services/form-recognizer/how-to-guides/build-a-custom-model.md
+11-11Lines changed: 11 additions & 11 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -43,19 +43,19 @@ Follow these tips to further optimize your data set for training:
43
43
44
44
## Upload your training data
45
45
46
-
Once you've put together the set of forms or documents for training, you'll need to upload it to an Azure blob storage container. If you don't know how to create an Azure storage account with a container, following the [Azure Storage quickstart for Azure portal](../../../storage/blobs/storage-quickstart-blobs-portal.md). You can use the free pricing tier (F0) to try the service, and upgrade later to a paid tier for production.
46
+
Once you've put together the set of forms or documents for training, you need to upload it to an Azure blob storage container. If you don't know how to create an Azure storage account with a container, following the [Azure Storage quickstart for Azure portal](../../../storage/blobs/storage-quickstart-blobs-portal.md). You can use the free pricing tier (F0) to try the service, and upgrade later to a paid tier for production.
47
47
48
48
## Video: Train your custom model
49
49
50
-
* Once you've gathered and uploaded your training dataset, you're ready to train your custom model. In the following video, we'll create a project and explore some of the fundamentals for successfully labeling and training a model.</br></br>
50
+
* Once you've gathered and uploaded your training dataset, you're ready to train your custom model. In the following video, we create a project and explore some of the fundamentals for successfully labeling and training a model.</br></br>
The Form Recognizer Studio provides and orchestrates all the API calls required to complete your dataset and train your model.
57
57
58
-
1. Start by navigating to the [Form Recognizer Studio](https://formrecognizer.appliedai.azure.com/studio). The first time you use the Studio, you'll need to [initialize your subscription, resource group, and resource](../quickstarts/try-v3-form-recognizer-studio.md). Then, follow the [prerequisites for custom projects](../quickstarts/try-v3-form-recognizer-studio.md#additional-prerequisites-for-custom-projects) to configure the Studio to access your training dataset.
58
+
1. Start by navigating to the [Form Recognizer Studio](https://formrecognizer.appliedai.azure.com/studio). The first time you use the Studio, you need to [initialize your subscription, resource group, and resource](../quickstarts/try-form-recognizer-studio.md). Then, follow the [prerequisites for custom projects](../quickstarts/try-form-recognizer-studio.md#added-prerequisites-for-custom-projects) to configure the Studio to access your training dataset.
59
59
60
60
1. In the Studio, select the **Custom models** tile, on the custom models page and select the **Create a project** button.
61
61
@@ -78,23 +78,23 @@ The Form Recognizer Studio provides and orchestrates all the API calls required
78
78
79
79
## Label your data
80
80
81
-
In your project, your first task is to label your dataset with the fields you wish to extract.
81
+
In your project, your first task is to label your dataset with the fields you wish to extract.
82
82
83
-
You'll see the files you uploaded to storage on the left of your screen, with the first file ready to be labeled.
83
+
The files you uploaded to storage are listed on the left of your screen, with the first file ready to be labeled.
84
84
85
85
1. To start labeling your dataset, create your first field by selecting the plus (➕) button on the top-right of the screen to select a field type.
86
86
87
87
:::image type="content" source="../media/how-to/studio-create-label.png" alt-text="Screenshot: Create a label.":::
88
88
89
89
1. Enter a name for the field.
90
90
91
-
1. To assign a value to the field, choose a word or words in the document and select the field in either the dropdown or the field list on the right navigation bar. You'll see the labeled value below the field name in the list of fields.
91
+
1. To assign a value to the field, choose a word or words in the document and select the field in either the dropdown or the field list on the right navigation bar. The labeled value is below the field name in the list of fields.
92
92
93
93
1. Repeat the process for all the fields you wish to label for your dataset.
94
94
95
95
1. Label the remaining documents in your dataset by selecting each document and selecting the text to be labeled.
96
96
97
-
You now have all the documents in your dataset labeled. If you look at the storage account, you'll find a *.labels.json* and *.ocr.json* files that correspond to each document in your training dataset and a new fields.json file. This training dataset will be submitted to train the model.
97
+
You now have all the documents in your dataset labeled. The *.labels.json* and *.ocr.json* files correspond to each document in your training dataset and a new fields.json file. This training dataset is submitted to train the model.
98
98
99
99
## Train your model
100
100
@@ -173,21 +173,21 @@ Follow these tips to further optimize your data set for training.
173
173
174
174
## Upload your training data
175
175
176
-
When you've put together the set of form documents that you'll use for training, you need to upload it to an Azure blob storage container. If you don't know how to create an Azure storage account with a container, follow the [Azure Storage quickstart for Azure portal](../../../storage/blobs/storage-quickstart-blobs-portal.md). Use the standard performance tier.
176
+
When you've put together the set of form documents for training, you need to upload it to an Azure blob storage container. If you don't know how to create an Azure storage account with a container, follow the [Azure Storage quickstart for Azure portal](../../../storage/blobs/storage-quickstart-blobs-portal.md). Use the standard performance tier.
177
177
178
-
If you want to use manually labeled data, you'll also have to upload the *.labels.json* and *.ocr.json* files that correspond to your training documents. You can use the [Sample Labeling tool](../label-tool.md) (or your own UI) to generate these files.
178
+
If you want to use manually labeled data, upload the *.labels.json* and *.ocr.json* files that correspond to your training documents. You can use the [Sample Labeling tool](../label-tool.md) (or your own UI) to generate these files.
179
179
180
180
### Organize your data in subfolders (optional)
181
181
182
-
By default, the [Train Custom Model](https://westus.dev.cognitive.microsoft.com/docs/services/form-recognizer-api-v2-1/operations/TrainCustomModelAsync) API will only use documents that are located at the root of your storage container. However, you can train with data in subfolders if you specify it in the API call. Normally, the body of the [Train Custom Model](https://westus.dev.cognitive.microsoft.com/docs/services/form-recognizer-api-v2-1/operations/TrainCustomModelAsync) call has the following format, where `<SAS URL>` is the Shared access signature URL of your container:
182
+
By default, the [Train Custom Model](https://westus.dev.cognitive.microsoft.com/docs/services/form-recognizer-api-v2-1/operations/TrainCustomModelAsync) API only uses documents that are located at the root of your storage container. However, you can train with data in subfolders if you specify it in the API call. Normally, the body of the [Train Custom Model](https://westus.dev.cognitive.microsoft.com/docs/services/form-recognizer-api-v2-1/operations/TrainCustomModelAsync) call has the following format, where `<SAS URL>` is the Shared access signature URL of your container:
183
183
184
184
```json
185
185
{
186
186
"source":"<SAS URL>"
187
187
}
188
188
```
189
189
190
-
If you add the following content to the request body, the API will train with documents located in subfolders. The `"prefix"` field is optional and will limit the training data set to files whose paths begin with the given string. So a value of `"Test"`, for example, will cause the API to look at only the files or folders that begin with the word "Test".
190
+
If you add the following content to the request body, the API trains with documents located in subfolders. The `"prefix"` field is optional and limits the training data set to files whose paths begin with the given string. So a value of `"Test"`, for example, causes the API to look at only the files or folders that begin with the word *Test*.
0 commit comments