You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/cognitive-services/language-service/custom-text-analytics-for-health/how-to/create-project.md
+5-5Lines changed: 5 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,7 +1,7 @@
1
1
---
2
2
title: Using Azure resources in custom Text Analytics for health
3
3
titleSuffix: Azure Cognitive Services
4
-
description: Learn about the steps for using Azure resources with custom TA4H.
4
+
description: Learn about the steps for using Azure resources with custom text analytics for health.
5
5
services: cognitive-services
6
6
author: aahill
7
7
manager: nitinme
@@ -15,17 +15,17 @@ ms.custom: language-service-custom-ta4h, references_regions, ignite-fall-2021, e
15
15
16
16
# How to create custom Text Analytics for health project
17
17
18
-
Use this article to learn how to set up the requirements for starting with custom TA4H and create a project.
18
+
Use this article to learn how to set up the requirements for starting with custom text analytics for health and create a project.
19
19
20
20
## Prerequisites
21
21
22
-
Before you start using custom TA4H, you will need:
22
+
Before you start using custom text analytics for health, you need:
23
23
24
24
* An Azure subscription - [Create one for free](https://azure.microsoft.com/free/cognitive-services).
25
25
26
26
## Create a Language resource
27
27
28
-
Before you start using custom TA4H, you will need an Azure Language resource. It is recommended to create your Language resource and connect a storage account to it in the Azure portal. Creating a resource in the Azure portal lets you create an Azure storage account at the same time, with all of the required permissions pre-configured. You can also read further in the article to learn how to use a pre-existing resource, and configure it to work with custom named entity recognition.
28
+
Before you start using custom text analytics for health, you'll need an Azure Language resource. It's recommended to create your Language resource and connect a storage account to it in the Azure portal. Creating a resource in the Azure portal lets you create an Azure storage account at the same time, with all of the required permissions preconfigured. You can also read further in the article to learn how to use a pre-existing resource, and configure it to work with custom named entity recognition.
29
29
30
30
You also will need an Azure storage account where you will upload your `.txt` documents that will be used to train a model to extract entities.
31
31
@@ -61,7 +61,7 @@ You can create a resource in the following ways:
61
61
62
62
## Create a custom Text Analytics for health project
63
63
64
-
Once your resource and storage container are configured, create a new custom TA4H project. A project is a work area for building your custom AI models based on your data. Your project can only be accessed by you and others who have access to the Azure resource being used. If you have labeled data, you can use it to get started by [importing a project](#import-project).
64
+
Once your resource and storage container are configured, create a new custom text analytics for health project. A project is a work area for building your custom AI models based on your data. Your project can only be accessed by you and others who have access to the Azure resource being used. If you have labeled data, you can use it to get started by [importing a project](#import-project).
Data labeling is a crucial step in development lifecycle. In this step, you label your documents with the new entities you defined in your schema to populate their learned components. This data will be used in the next step when training your model so that your model can learn from the labeled data to to know which entities to extract. If you already have labeled data, you can directly [import](create-project.md#import-project) it into your project but you need to make sure that your data follows the [accepted data format](../concepts/data-formats.md). See [create project](create-project.md#import-project) to learn more about importing labeled data into your project. If your data isn't labeled already, you can label it in the [Language Studio](https://aka.ms/languageStudio).
18
+
Data labeling is a crucial step in development lifecycle. In this step, you label your documents with the new entities you defined in your schema to populate their learned components. This data will be used in the next step when training your model so that your model can learn from the labeled data to know which entities to extract. If you already have labeled data, you can directly [import](create-project.md#import-project) it into your project, but you need to make sure that your data follows the [accepted data format](../concepts/data-formats.md). See [create project](create-project.md#import-project) to learn more about importing labeled data into your project. If your data isn't labeled already, you can label it in the [Language Studio](https://aka.ms/languageStudio).
19
19
20
20
## Prerequisites
21
21
@@ -28,11 +28,11 @@ See the [project development lifecycle](../overview.md#project-development-lifec
28
28
29
29
## Data labeling guidelines
30
30
31
-
After preparing your data, designing your schema and creating your project, you will need to label your data. Labeling your data is important so your model knows which words will be associated with the entity types you need to extract. When you label your data in [Language Studio](https://aka.ms/languageStudio) (or import labeled data), these labels will be stored in the JSON document in your storage container that you have connected to this project.
31
+
After preparing your data, designing your schema and creating your project, you will need to label your data. Labeling your data is important so your model knows which words will be associated with the entity types you need to extract. When you label your data in [Language Studio](https://aka.ms/languageStudio) (or import labeled data), these labels are stored in the JSON document in your storage container that you have connected to this project.
32
32
33
33
As you label your data, keep in mind:
34
34
35
-
* You cannot add labels for Text Analytics for health entities as they are pretrained prebuilt entities. You can only add labels to new entity categories that you defined during schema definition.
35
+
* You can't add labels for Text Analytics for health entities as they're pretrained prebuilt entities. You can only add labels to new entity categories that you defined during schema definition.
36
36
37
37
<!--If you want to improve the recall for a prebuilt entity, you can extend it by adding a list component while you are [defining your schema](design-schema.md).-->
38
38
@@ -81,7 +81,7 @@ Use the following steps to label your data:
81
81
82
82
:::image type="content" source="../media/tag-options.png" alt-text="A screenshot showing the labeling options offered in Custom NER." lightbox="../media/tag-options.png":::
83
83
84
-
6. In the right side pane under the **Labels** pivot you can find all the entity types in your project and the count of labeled instances per each. Note that the prebuilt entities will be showing for reference but you will not be able to label for these prebuilt entities as they are pretrained.
84
+
6. In the right side pane under the **Labels** pivot you can find all the entity types in your project and the count of labeled instances per each. The prebuilt entities will be shown for reference but you will not be able to label for these prebuilt entities as they are pretrained.
85
85
86
86
7. In the bottom section of the right side pane you can add the current document you are viewing to the training set or the testing set. By default all the documents are added to your training set. <!--Learn more about [training and testing sets](train-model.md#data-splitting) and how they are used for model training and evaluation.-->
0 commit comments