You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
> **Instance Segmentation** projects cannot use consensus labeling.
132
+
126
133
## Use ML-assisted data labeling
127
134
128
135
The **ML-assisted labeling** page lets you trigger automatic machine learning models to accelerate labeling tasks. Medical images (".dcm") are not included in assisted labeling.
@@ -131,14 +138,15 @@ At the beginning of your labeling project, the items are shuffled into a random
131
138
132
139
Select *Enable ML assisted labeling* and specify a GPU to enable assisted labeling. If you don't have one in your workspace, a GPU cluster will be created for you and added to your workspace. The cluster is created with a minimum of 0 nodes, which means it doesn't cost anything when it's not in use.
133
140
134
-
135
141
ML-assisted labeling consists of two phases:
136
142
137
143
* Clustering
138
144
* Prelabeling
139
145
140
146
The exact number of labeled data necessary to start assisted labeling is not a fixed number. This can vary significantly from one labeling project to another. For some projects, is sometimes possible to see prelabel or cluster tasks after 300 items have been manually labeled. ML Assisted Labeling uses a technique called *Transfer Learning*, which uses a pre-trained model to jump-start the training process. If your dataset's classes are similar to those in the pre-trained model, pre-labels may be available after only a few hundred manually labeled items. If your dataset is significantly different from the data used to pre-train the model, it may take much longer.
141
147
148
+
When you're using consensus labeling, the consensus label is used for training.
149
+
142
150
Since the final labels still rely on input from the labeler, this technology is sometimes called *human in the loop* labeling.
143
151
144
152
> [!NOTE]
@@ -192,6 +200,24 @@ On the right side is a distribution of the labels for those tasks that are compl
192
200
193
201
On the **Data** tab, you can see your dataset and review labeled data. Scroll through the labeled data to see the labels. If you see incorrectly labeled data, select it and choose **Reject**, which will remove the labels and put the data back into the unlabeled queue.
194
202
203
+
If your project uses consensus labeling, you'll also want to review those images without a consensus. To do so:
1. Under **Labeled datapoints**, select **Consensus labels in need of review**. This shows only those images where a consensus was not achieved among the labelers.
212
+
213
+
:::image type="content" source="media/how-to-create-labeling-projects/select-need-review.png" alt-text="Screenshot: Select labels in need of review.":::
214
+
215
+
1. For each image in need of review, select the **Consensus label** dropdown to view the conflicting labels.
1. While you can select an individual to see just their label(s), you can only update or reject the labels from the top choice, **Consensus label (preview)**.
220
+
195
221
### Details tab
196
222
197
223
View and change details of your project. In this tab you can:
@@ -213,7 +239,7 @@ View and change details of your project. In this tab you can:
213
239
214
240
## Export the labels
215
241
216
-
Use the **Export** button on the **Project details** page of your labeling project. You can export the label data for Machine Learning experimentation at any time.
242
+
Use the **Export** button on the **Project details** page of your labeling project. You can export the label data for Machine Learning experimentation at any time.
217
243
218
244
* Image labels can be exported as:
219
245
*[COCO format](http://cocodataset.org/#format-data).The COCO file is created in the default blob store of the Azure Machine Learning workspace in a folder within *Labeling/export/coco*.
The **ML-assisted labeling** page lets you trigger automatic machine learning models to accelerate labeling tasks. ML-assisted labeling is available for both file (.txt) and tabular (.csv) text data inputs.
@@ -140,7 +144,9 @@ At the beginning of your labeling project, the items are shuffled into a random
140
144
141
145
For training the text DNN model used by ML-assist, the input text per training example will be limited to approximately the first 128 words in the document. For tabular input, all text columns are first concatenated before applying this limit. This is a practical limit imposed to allow for the model training to complete in a timely manner. The actual text in a document (for file input) or set of text columns (for tabular input) can exceed 128 words. The limit only pertains to what is internally leveraged by the model during the training process.
142
146
143
-
The exact number of labeled items necessary to start assisted labeling isn't a fixed number. This can vary significantly from one labeling project to another, depending on many factors, including the number of labels classes and label distribution.
147
+
The exact number of labeled items necessary to start assisted labeling isn't a fixed number. This can vary significantly from one labeling project to another, depending on many factors, including the number of labels classes and label distribution.
148
+
149
+
When you're using consensus labeling, the consensus label is used for training.
144
150
145
151
Since the final labels still rely on input from the labeler, this technology is sometimes called *human in the loop* labeling.
146
152
@@ -179,6 +185,24 @@ On the right side is a distribution of the labels for those tasks that are compl
179
185
180
186
On the **Data** tab, you can see your dataset and review labeled data. Scroll through the labeled data to see the labels. If you see incorrectly labeled data, select it and choose **Reject**, which will remove the labels and put the data back into the unlabeled queue.
181
187
188
+
If your project uses consensus labeling, you'll also want to review those images without a consensus. To do so:
1. Under **Labeled datapoints**, select **Consensus labels in need of review**. This shows only those images where a consensus was not achieved among the labelers.
197
+
198
+
:::image type="content" source="media/how-to-create-labeling-projects/select-need-review.png" alt-text="Screenshot: Select labels in need of review.":::
199
+
200
+
1. For each item in need of review, select the **Consensus label** dropdown to view the conflicting labels.
1. While you can select an individual to see just their label(s), you can only update or reject the labels from the top choice, **Consensus label (preview)**.
205
+
182
206
### Details tab
183
207
184
208
View and change details of your project. In this tab you can:
To get more accurate labels, use the **Quality control** page to send each item to multiple labelers.
10
+
11
+
> [!IMPORTANT]
12
+
> Consensus labeling is currently in public preview.
13
+
> The preview version is provided without a service level agreement, and it's not recommended for production workloads. Certain features might not be supported or might have constrained capabilities.
14
+
> For more information, see [Supplemental Terms of Use for Microsoft Azure Previews](https://azure.microsoft.com/support/legal/preview-supplemental-terms/).
15
+
16
+
Select **Enable consensus labeling (preview)** to have each item sent to multiple labelers. Then set the **Minimum labelers** and **Maximum labelers** to specify how many labelers to use. Make sure you have as many labelers available as your maximum number. You can't later change these settings once the project has started.
17
+
18
+
If a consensus is reached from the minimum number of labelers, the item is labeled. If a consensus isn't reached, the item will be sent to more labelers. If there's no consensus after the item goes to the maximum number of labelers, its status will be `Needs Review`, and the project owner will be responsible for labeling the item.
0 commit comments