You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/cognitive-services/form-recognizer/quickstarts/label-tool.md
+11-7Lines changed: 11 additions & 7 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -8,7 +8,7 @@ manager: nitinme
8
8
ms.service: cognitive-services
9
9
ms.subservice: forms-recognizer
10
10
ms.topic: quickstart
11
-
ms.date: 11/14/2019
11
+
ms.date: 02/19/2020
12
12
ms.author: pafarley
13
13
---
14
14
@@ -24,10 +24,14 @@ To complete this quickstart, you must have:
24
24
25
25
- A set of at least six forms of the same type. You'll use this data to train the model and test a form. You can use a [sample data set](https://go.microsoft.com/fwlink/?linkid=2090451) for this quickstart. Upload the training files to the root of a blob storage container in an Azure Storage account.
You'll use the Docker engine to run the sample labeling tool. Follow these steps to set up the Docker container. For a primer on Docker and container basics, see the [Docker overview](https://docs.docker.com/engine/docker-overview/).
30
-
1. First, install Docker on a host computer. The host computer can be your local computer ([Windows](https://docs.docker.com/docker-for-windows/), [MacOS](https://docs.docker.com/docker-for-mac/), or [Linux](https://docs.docker.com/install/)). Or, you can use a Docker hosting service in Azure, such as the [Azure Kubernetes Service](https://docs.microsoft.com/azure/aks/index), [Azure Container Instances](https://docs.microsoft.com/azure/container-instances/index), or a Kubernetes cluster [deployed to an Azure Stack](https://docs.microsoft.com/azure-stack/user/azure-stack-solution-template-kubernetes-deploy?view=azs-1910). The host computer must meet the following hardware requirements:
34
+
1. First, install Docker on a host computer. The host computer can be your local computer ([Windows](https://docs.docker.com/docker-for-windows/), [macOS](https://docs.docker.com/docker-for-mac/), or [Linux](https://docs.docker.com/install/)). Or, you can use a Docker hosting service in Azure, such as the [Azure Kubernetes Service](https://docs.microsoft.com/azure/aks/index), [Azure Container Instances](https://docs.microsoft.com/azure/container-instances/index), or a Kubernetes cluster [deployed to an Azure Stack](https://docs.microsoft.com/azure-stack/user/azure-stack-solution-template-kubernetes-deploy?view=azs-1910). The host computer must meet the following hardware requirements:
31
35
32
36
| Container | Minimum | Recommended|
33
37
|:--|:--|:--|
@@ -66,7 +70,7 @@ Enable CORS on your storage account. Select your storage account in the Azure po
66
70
67
71
## Connect to the sample labeling tool
68
72
69
-
The sample labeling tool connects to a source (where your original forms are) and a target (the location where it exports the created labels and output data).
73
+
The sample labeling tool connects to a source (where your original forms are) and a target (where it exports the created labels and output data).
70
74
71
75
Connections can be set up and shared across projects. They use an extensible provider model, so you can easily add new source/target providers.
72
76
@@ -85,7 +89,7 @@ Fill in the fields with the following values:
85
89
In the sample labeling tool, projects store your configurations and settings. Create a new project and fill in the fields with the following values:
86
90
87
91
* **Display Name** - the project display name
88
-
* **Security Token** - Some project settings can include sensitive values, such as API keys or other shared secrets. Each project will generate a security token that can be used to encrypt/decrypt sensitive project settings. Security tokens can be found in Application Settings by clicking the gear icon in the lower corner of the left navigation bar.
92
+
* **Security Token** - Some project settings can include sensitive values, such as API keys or other shared secrets. Each project will generate a security token that can be used to encrypt/decrypt sensitive project settings. You can find security tokens in the Application Settings by clicking the gear icon in the lower corner of the left navigation bar.
89
93
* **Source Connection** - The Azure Blob Storage connection you created in the previous step that you would like to use for this project.
90
94
* **Folder Path** - Optional - If your source forms are located in a folder on the blob container, specify the folder name here
91
95
* **Form Recognizer Service Uri** - Your Form Recognizer endpoint URL.
@@ -142,7 +146,7 @@ After training finishes, examine the **Average Accuracy** value. If it's low, yo
142
146
143
147
## Analyze a form
144
148
145
-
Click on the Predict (rectangles) icon on the left to test your model. Upload a form document that you didn't use in the training process. Then click the **Predict** button on the right to get key/value predictions for the form. The tool will apply tags in bounding boxes and will report the confidence of each tag.
149
+
Click on the Predict (rectangles) icon on the left to test your model. Upload a form document that you haven't used in the training process. Then click the **Predict** button on the right to get key/value predictions for the form. The tool will apply tags in bounding boxes and will report the confidence of each tag.
146
150
147
151
> [!TIP]
148
152
> You can also run the Analyze API with a REST call. To learn how to do this, see [Train with labels using Python](./python-labeled-data.md).
@@ -151,7 +155,7 @@ Click on the Predict (rectangles) icon on the left to test your model. Upload a
151
155
152
156
Depending on the reported accuracy, you may want to do further training to improve the model. After you've done a prediction, examine the confidence values for each of the applied tags. If the average accuracy training value was high, but the confidence scores are low (or the results are inaccurate), you should add the file used for prediction into the training set, label it, and train again.
153
157
154
-
The reported average accuracy, confidence scores, and actual accuracy can be inconsistent when the documents being analyzed are different from those used in training. Keep in mind that some documents look similar when viewed by people but can look distinct to the AI model. For example, you might train with a form type that has two variations, where the training set consists of 20% variation A and 80% variation B. During prediction, the confidence scores for documents of variation A are likely to be lower.
158
+
The reported average accuracy, confidence scores, and actual accuracy can be inconsistent when the analyzed documents differ from those used in training. Keep in mind that some documents look similar when viewed by people but can look distinct to the AI model. For example, you might train with a form type that has two variations, where the training set consists of 20% variation A and 80% variation B. During prediction, the confidence scores for documents of variation A are likely to be lower.
155
159
156
160
## Save a project and resume later
157
161
@@ -161,7 +165,7 @@ To resume your project at another time or in another browser, you need to save y
161
165
Go to your project settings page (slider icon) and take note of the security token name. Then go to your application settings (gear icon), which shows all of the security tokens in your current browser instance. Find your project's security token and copy its name and key value to a secure location.
162
166
163
167
### Restore project credentials
164
-
When you want to resume your project, you first need to create a connection to the same blob storage container. Follow the steps above to do this. Then, go to the application settings page (gear icon) and see if your project's security token is there. If it isn't, add a new security token and copy over your token name and key from the previous step. Then click Save Settings.
168
+
When you want to resume your project, you first need to create a connection to the same blob storage container. Repeat the steps above to do this. Then, go to the application settings page (gear icon) and see if your project's security token is there. If it isn't, add a new security token and copy over your token name and key from the previous step. Then click Save Settings.
165
169
166
170
### Resume a project
167
171
Finally, go to the main page (house icon) and click Open Cloud Project. Then select the blob storage connection, and select your project's *.vott* file. The application will load all of the project's settings because it has the security token.
Copy file name to clipboardExpand all lines: articles/cognitive-services/form-recognizer/quickstarts/python-labeled-data.md
+6-2Lines changed: 6 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -8,7 +8,7 @@ manager: nitinme
8
8
ms.service: cognitive-services
9
9
ms.subservice: forms-recognizer
10
10
ms.topic: quickstart
11
-
ms.date: 01/27/2020
11
+
ms.date: 02/19/2020
12
12
ms.author: pafarley
13
13
14
14
---
@@ -25,9 +25,13 @@ To complete this quickstart, you must have:
25
25
-[Python](https://www.python.org/downloads/) installed (if you want to run the sample locally).
26
26
- A set of at least six forms of the same type. You will use this data to train the model and test a form. You can use a [sample data set](https://go.microsoft.com/fwlink/?linkid=2090451) for this quickstart. Upload the training files to the root of a blob storage container in an Azure Storage account.
First you'll need to set up the required input data. The labeled data feature has special input requirements beyond those needed to train a custom model.
34
+
Next you'll need to set up the required input data. The labeled data feature has special input requirements beyond those needed to train a custom model.
31
35
32
36
Make sure all the training documents are of the same format. If you have forms in multiple formats, organize them into sub-folders based on common format. When you train, you'll need to direct the API to a sub-folder.
After you've called the **Analyze Layout** API, you call the **[Get Analyze Layout Result](https://westus2.dev.cognitive.microsoft.com/docs/services/form-recognizer-api-v2-preview/operations/GetAnalyzeLayoutResult)** API to get the status of the operation and the extracted data. Add the following code to the bottom of your Python script. This uses the operation ID value in a new API call. This script calls the API at regular intervals until the results are available. We recommend an interval of one second or more.
84
+
After you've called the **Analyze Layout** API, you call the **[Get Analyze Layout Result](https://westus2.dev.cognitive.microsoft.com/docs/services/form-recognizer-api-v2-preview/operations/GetAnalyzeLayoutResult)** API to get the status of the operation and the extracted data. Add the following code to the bottom of your Python script. This code uses the operation ID value in a new API call. This script calls the API at regular intervals until the results are available. We recommend an interval of one second or more.
0 commit comments