You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-services/computer-vision/overview-ocr.md
+14-14Lines changed: 14 additions & 14 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -17,29 +17,29 @@ ms.custom: devx-track-csharp
17
17
# OCR - Optical Character Recognition
18
18
19
19
> [!WARNING]
20
-
> This service, including the Azure AI Vision legacy [OCR API in v3.2](/rest/api/computervision/recognize-printed-text?view=rest-computervision-v3.2) and [RecognizeText API in v2.1](/rest/api/computervision/recognize-printed-text/recognize-printed-text?view=rest-computervision-v2.1), is not recommended for use.
20
+
> We don't recommend using this service, including the Azure AI Vision legacy [OCR API in v3.2](/rest/api/computervision/recognize-printed-text?view=rest-computervision-v3.2) and [RecognizeText API in v2.1](/rest/api/computervision/recognize-printed-text/recognize-printed-text?view=rest-computervision-v2.1).
OCR or Optical Character Recognition is also referred to as text recognition or text extraction. Machine-learning-based OCR techniques allow you to extract printed or handwritten text from images such as posters, street signs and product labels, as well as from documents like articles, reports, forms, and invoices. The text is typically extracted as words, text lines, and paragraphs or text blocks, enabling access to digital version of the scanned text. This eliminates or significantly reduces the need for manual data entry.
25
+
OCR or Optical Character Recognition is also referred to as text recognition or text extraction. Machine-learning-based OCR techniques allow you to extract printed or handwritten text from images such as posters, street signs, and product labels, as well as from documents like articles, reports, forms, and invoices. The text is typically extracted as words, text lines, and paragraphs or text blocks, enabling access to digital version of the scanned text. This capability eliminates or significantly reduces the need for manual data entry.
26
26
27
27
28
28
29
29
## OCR engine
30
30
31
-
Microsoft's **Read** OCR engine is composed of multiple advanced machine-learning based models supporting [global languages](./language-support.md). It can extract printed and handwritten text including mixed languages and writing styles. **Read**is available as cloud service and on-premises container for deployment flexibility. It's also available as a synchronous API for single, non-document, image-only scenarios with performance enhancements that make it easier to implement OCR-assisted user experiences.
31
+
Microsoft's **Read** OCR engine uses multiple advanced machine-learning models that support [global languages](./language-support.md). It extracts printed and handwritten text, including mixed languages and writing styles. You can use **Read** as a cloud service or as an on-premises container for flexible deployment. It's also available as a synchronous API for single, non-document, image-only scenarios with performance enhancements that simplify implementing OCR-assisted user experiences.
32
32
33
33
34
34
35
35
36
36
## How is OCR related to Intelligent Document Processing (IDP)?
37
37
38
-
Intelligent Document Processing (IDP) uses OCR as its foundational technology to additionally extract structure, relationships, key-values, entities, and other document-centric insights with an advanced machine-learning based AI service like [Document Intelligence](../../ai-services/document-intelligence/overview.md). Document Intelligence includes a document-optimized version of **Read** as its OCR engine while delegating to other models for higher-end insights. If you are extracting text from scanned and digital documents, use [Document Intelligence Read OCR](../document-intelligence/prebuilt/read.md).
38
+
Intelligent Document Processing (IDP) uses OCR as its foundational technology to extract structure, relationships, key-values, entities, and other document-centric insights with an advanced machine-learning based AI service like [Document Intelligence](../../ai-services/document-intelligence/overview.md). Document Intelligence includes a document-optimized version of **Read** as its OCR engine while delegating to other models for higher-end insights. If you're extracting text from scanned and digital documents, use [Document Intelligence Read OCR](../document-intelligence/prebuilt/read.md).
39
39
40
40
## How to use OCR
41
41
42
-
Try out OCR by using Vision Studio. Then follow one of the links to the Read edition that best meet your requirements.
42
+
Try out OCR by using Vision Studio. Then follow one of the links to the Read edition that best meets your requirements.
@@ -48,7 +48,7 @@ Try out OCR by using Vision Studio. Then follow one of the links to the Read edi
48
48
49
49
## OCR supported languages
50
50
51
-
Both **Read** versions available today in Azure AI Vision support several languages for printed and handwritten text. OCR for printed text includes support for English, French, German, Italian, Portuguese, Spanish, Chinese, Japanese, Korean, Russian, Arabic, Hindi, and other international languages that use Latin, Cyrillic, Arabic, and Devanagari scripts. OCR for handwritten text includes support for English, Chinese Simplified, French, German, Italian, Japanese, Korean, Portuguese, and Spanish languages.
51
+
Both **Read** versions available today in Azure AI Vision support several languages for printed and handwritten text. OCR for printed text supports English, French, German, Italian, Portuguese, Spanish, Chinese, Japanese, Korean, Russian, Arabic, Hindi, and other international languages that use Latin, Cyrillic, Arabic, and Devanagari scripts. OCR for handwritten text supports English, Chinese Simplified, French, German, Italian, Japanese, Korean, Portuguese, and Spanish languages.
52
52
53
53
Refer to the full list of [OCR-supported languages](./language-support.md#optical-character-recognition-ocr).
54
54
@@ -57,24 +57,24 @@ Refer to the full list of [OCR-supported languages](./language-support.md#optica
57
57
The Read OCR model is available in Azure AI Vision and Document Intelligence with common baseline capabilities while optimizing for respective scenarios. The following list summarizes the common features:
58
58
59
59
* Printed and handwritten text extraction in supported languages
60
-
* Pages, text lines and words with location and confidence scores
60
+
* Pages, text lines, and words with location and confidence scores
61
61
* Support for mixed languages, mixed mode (print and handwritten)
62
62
* Available as Distroless Docker container for on-premises deployment
63
63
64
64
## Use the OCR cloud APIs or deploy on-premises
65
65
66
-
The cloud APIs are the preferred option for most customers because of their ease of integration and fast productivity out of the box. Azure and the Azure AI Vision service handle scale, performance, data security, and compliance needs while you focus on meeting your customers' needs.
66
+
Most customers prefer the cloud APIs because they're easy to integrate and offer fast productivity out of the box. Azure and the Azure AI Vision service handle scale, performance, data security, and compliance needs while you focus on meeting your customers' needs.
67
67
68
68
For on-premises deployment, the [Read Docker container](./computer-vision-how-to-install-containers.md) enables you to deploy the Azure AI Vision v3.2 generally available OCR capabilities in your own local environment. Containers are great for specific security and data governance requirements.
69
69
70
70
71
71
## Input requirements
72
72
73
-
The **Read** API takes images and documents as its input. The images and documents must meet the following requirements:
73
+
The **Read** API takes images and documents as input. The images and documents must meet the following requirements:
74
74
75
75
* Supported file formats are JPEG, PNG, BMP, PDF, and TIFF.
76
-
* For PDF and TIFF files, up to 2,000 pages (only the first two pages for the free tier) are processed.
77
-
* The file size of images must be less than 500 MB (4 MB for the free tier) with dimensions at least 50 x 50 pixels and at most 10,000 x 10,000 pixels. PDF files don't have a size limit.
76
+
* For PDF and TIFF files, up to 2,000 pages are processed (only the first two pages for the free tier).
77
+
* The file size of images must be less than 500 MB (4 MB for the free tier) with dimensions of at least 50 x 50 pixels and at most 10,000 x 10,000 pixels. PDF files don't have a size limit.
78
78
* The minimum height of the text to be extracted is 12 pixels for a 1024 x 768 image, which corresponds to about 8-point font text at 150 DPI.
79
79
80
80
>[!NOTE]
@@ -86,6 +86,6 @@ As with all of the Azure AI services, developers using the Azure AI Vision servi
86
86
87
87
## Next steps
88
88
89
-
- OCR for general (non-document) images: try the [Azure AI Vision 4.0 preview Image Analysis REST API quickstart](./concept-ocr.md).
90
-
- OCR for PDF, Office and HTML documents and document images: start with [Document Intelligence Read](../../ai-services/document-intelligence/concept-read.md).
91
-
-Looking for the previous GA version? Refer to the [Azure AI Vision 3.2 GA SDK or REST API quickstarts](./quickstarts-sdk/client-library.md).
89
+
-For OCR with general (non-document) images, try the [Azure AI Vision 4.0 preview Image Analysis REST API quickstart](./concept-ocr.md).
90
+
-For OCR with PDF, Office, and HTML documents, as well as document images, start with [Document Intelligence Read](../../ai-services/document-intelligence/concept-read.md).
91
+
-For the previous GA version, see the [Azure AI Vision 3.2 GA SDK or REST API quickstarts](./quickstarts-sdk/client-library.md).
Copy file name to clipboardExpand all lines: articles/ai-services/content-understanding/quickstart/use-ai-foundry.md
+13-13Lines changed: 13 additions & 13 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -18,7 +18,7 @@ This quickstart shows you how to use the Content Understanding service in the [*
18
18
19
19
Suppose you have files—such as documents, images, audio, or video—and you want to automatically extract key information from them. With Content Understanding, you can create a task to organize your data processing, define a field schema that specifies the information to extract or generate, and then build an analyzer. The analyzer becomes an API endpoint that you can integrate into your applications or workflows.
20
20
21
-
In this guide, we walk you through building and testing an analyzer for your scenario. You can start from scratch or use suggested templates for common use cases.
21
+
In this guide, you build and test an analyzer for your scenario. You can start from scratch or use suggested templates for common use cases.
22
22
23
23
:::image type="content" source="../media/overview/component-overview-updated.png" alt-text="Screenshot of Content Understanding overview, process, and workflow." lightbox="../media/overview/component-overview-updated.png" :::
24
24
@@ -28,27 +28,27 @@ To get started, make sure you have the following resources and permissions:
28
28
29
29
* An Azure subscription. If you don't have an Azure subscription, [create a free account](https://azure.microsoft.com/free/).
30
30
31
-
* An [Azure AI Foundry hub-based project](../../../ai-foundry/how-to/create-projects.md) created in one of the following [supported regions](../service-limits.md): `westus`, `swedencentral`, or `australiaeast`. A project is used to organize your work and save state while building customized AI apps. You can create a project from the [home page of AI Foundry](https://aka.ms/foundry-home-page), or the [Content Understanding landing page](https://aka.ms/cu-landing).
31
+
* An [Azure AI Foundry hub-based project](../../../ai-foundry/how-to/create-projects.md) created in one of the following [supported regions](../service-limits.md): `westus`, `swedencentral`, or `australiaeast`. Use a project to organize your work and save state while building customized AI apps. You can create a project from the [home page of AI Foundry](https://aka.ms/foundry-home-page), or the [Content Understanding landing page](https://aka.ms/cu-landing).
32
32
33
33
[!INCLUDE [hub based project required](../../../ai-foundry/includes/uses-hub-only.md)]
34
34
35
35
## Create your single-file task powered by Content Understanding Standard mode
36
36
37
-
Follow these steps to create a custom task in the Azure AI Foundry. This task is used to build your first analyzer.
37
+
Follow these steps to create a custom task in the Azure AI Foundry. Use this task to build your first analyzer.
38
38
39
39
1. Go to the **Home** page of [Azure AI Foundry](https://ai.azure.com/?cid=learnDocs).
40
40
1. Select your hub based project. You might need to select **View all resources** to see your project.
41
41
1. Select **Content Understanding** from the left navigation pane.
42
42
1. Select **+ Create**.
43
-
2. In this guide, you create a `Single-file task` utilizing Content Understanding Standard mode, but if you're interested in creating a multi-file task utilizing Pro mode, refer to [Create an Azure AI Content Understanding multi-file task in the Azure AI Foundry portal](./use-ai-foundry-pro-mode.md). For more information on which mode is right for your scenario, check out [Azure AI Content Understanding pro and standard modes](../concepts/standard-pro-modes.md).
43
+
1. In this guide, you create a `Single-file task` utilizing Content Understanding Standard mode, but if you're interested in creating a multi-file task utilizing Pro mode, refer to [Create an Azure AI Content Understanding multi-file task in the Azure AI Foundry portal](./use-ai-foundry-pro-mode.md). For more information on which mode is right for your scenario, check out [Azure AI Content Understanding pro and standard modes](../concepts/standard-pro-modes.md).
44
44
1. Enter a name for your task. Optionally, enter a description and change other settings.
45
45
1. Select **Create**.
46
46
47
47
## Create your first analyzer
48
48
49
-
Now that everything is configured to get started, we can walk through how to build your first analyzer.
49
+
Now that everything is configured, you can build your first analyzer.
50
50
51
-
When you create a single-file Content Understanding task, you start by uploading a sample of your data and building your field schema. The schema is the customizable framework that allows the analyzer to extract insights from your data. In this example, the schema is created to extract key data from an invoice document, but you can bring in any type of data and the steps remain the same. For a complete list of supported file types, see [input file limits](../service-limits.md#input-file-limits).
51
+
When you create a single-file Content Understanding task, you start by uploading a sample of your data and building your field schema. The schema is the customizable framework that allows the analyzer to extract insights from your data. In this example, you create the schema to extract key data from an invoice document, but you can bring in any type of data and the steps remain the same. For a complete list of supported file types, see [input file limits](../service-limits.md#input-file-limits).
52
52
53
53
1. Upload a [sample file of an invoice document](https://github.com/Azure-Samples/azure-ai-content-understanding-python/raw/refs/heads/main/data/invoice.pdf) or any other data relevant to your scenario.
54
54
@@ -60,7 +60,7 @@ When you create a single-file Content Understanding task, you start by uploading
60
60
61
61
:::image type="content" source="../media/quickstarts/invioce-template.png" alt-text="Screenshot of analyzer templates.":::
62
62
63
-
1. Next, you can add fields to your schema to reflect all of the outputs you want to generate.
63
+
1. Next, add fields to your schema to reflect all of the outputs you want to generate.
64
64
65
65
* Specify clear and simple field names. Some example fields might include **vendorName**, **items**, **price**.
66
66
@@ -70,29 +70,29 @@ When you create a single-file Content Understanding task, you start by uploading
70
70
71
71
* Specify the method to generate the value for each field.
72
72
73
-
For best practices on how to define your field schema, refer to [best practices for Azure AI Content Understanding](../concepts//best-practices.md). It may take a few minutes to build out your schema.
73
+
For best practices on how to define your field schema, refer to [best practices for Azure AI Content Understanding](../concepts//best-practices.md). It might take a few minutes to build out your schema.
74
74
75
-
1.Once you feel that the schema is ready to test, select **Save**. You can always come back and make changes if needed.
75
+
1.When your schema is ready to test, select **Save**. You can always come back and make changes if needed.
76
76
77
77
:::image type="content" source="../media/quickstarts/define-invoice-schema.png" alt-text="Screenshot of completed schema.":::
78
78
79
79
1. With the completed schema, Content Understanding now generates the output on your sample data. At this step, you can add more data to test the analyzer's accuracy or make changes to the schema if needed.
80
80
81
81
:::image type="content" source="../media/quickstarts/test-invoice.png" alt-text="Screenshot of schema testing step.":::
82
82
83
-
1.Once you're satisfied with the quality of your output, select **Build analyzer**. This action creates an analyzer ID that you can integrate into your own applications, allowing you to call the analyzer from your code.
83
+
1.When you're satisfied with the quality of your output, select **Build analyzer**. This action creates an analyzer ID that you can integrate into your own applications, allowing you to call the analyzer from your code.
84
84
85
85
:::image type="content" source="../media/quickstarts/build-invoice-analyzer.png" alt-text="Screenshot of built analyzer.":::
86
86
87
-
Now you successfully built your first Content Understanding analyzer, and are ready to start extracting insights from your data. Check out [Quickstart: Azure AI Content Understanding REST APIs](./use-rest-api.md) to utilize the REST API to call your analyzer.
87
+
You've successfully built your first Content Understanding analyzer and are ready to start extracting insights from your data. Check out [Quickstart: Azure AI Content Understanding REST APIs](./use-rest-api.md) to utilize the REST API to call your analyzer.
88
88
89
89
## Sharing your project
90
90
91
-
In order to share and manage access to the project you created, navigate to the Management Center, found at the bottom of the navigation for your project:
91
+
To share the project you created and manage access, go to the Management Center. You can find it at the bottom of the navigation pane for your project:
92
92
93
93
:::image type="content" source="../media/quickstarts/cu-landing-page.png" alt-text="Screenshot of where to find management center.":::
94
94
95
-
You can manage the users and their individual roles here:
95
+
In the Management Center, you can manage users and assign individual roles:
96
96
97
97
:::image type="content" source="../media/quickstarts/management-center.png" alt-text="Screenshot of Project users section of management center.":::
0 commit comments