You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-foundry/concepts/models-featured.md
+2-1Lines changed: 2 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -250,7 +250,7 @@ See [the Microsoft model collection in Azure AI Foundry portal](https://ai.azure
250
250
251
251
Mistral AI offers two categories of models, namely:
252
252
253
-
-_Premium models_: These include Mistral Large, Mistral Small, Mistral-OCR-2503, and Ministral 3B models, and are available as serverless APIs with pay-as-you-go token-based billing.
253
+
-_Premium models_: These include Mistral Large, Mistral Small, Mistral-OCR-2503, Mistral Medium 3 (25.05), and Ministral 3B models, and are available as serverless APIs with pay-as-you-go token-based billing.
254
254
-_Open models_: These include Mistral-small-2503, Codestral, and Mistral Nemo (that are available as serverless APIs with pay-as-you-go token-based billing), and [Mixtral-8x7B-Instruct-v01, Mixtral-8x7B-v01, Mistral-7B-Instruct-v01, and Mistral-7B-v01](../how-to/deploy-models-mistral-open.md)(that are available to download and run on self-hosted managed endpoints).
255
255
256
256
@@ -262,6 +262,7 @@ Mistral AI offers two categories of models, namely:
|[Mistral-medium-2505](https://aka.ms/aistudio/landing/mistral-medium-2505)|[chat-completion](../model-inference/how-to/use-chat-completions.md?context=/azure/ai-foundry/context/context)| - **Input:** text (128,000 tokens), image <br /> - **Output:** text (128,000 tokens) <br /> - **Tool calling:** No <br /> - **Response formats:** Text, JSON |
265
266
|[Mistral-OCR-2503](https://aka.ms/aistudio/landing/mistral-ocr-2503)|[image to text](../how-to/use-image-models.md)| - **Input:** image or PDF pages (1,000 pages, max 50MB PDF file) <br> - **Output:** text <br /> - **Tool calling:** No <br /> - **Response formats:** Text, JSON, Markdown |
266
267
|[Mistral-small-2503](https://aka.ms/aistudio/landing/mistral-small-2503)|[chat-completion (with images)](../model-inference/how-to/use-chat-multi-modal.md?context=/azure/ai-foundry/context/context)| - **Input:** text and images (131,072 tokens), <br> image-based tokens are 16px x 16px <br> blocks of the original images <br /> - **Output:** text (4,096 tokens) <br /> - **Tool calling:** Yes <br /> - **Response formats:** Text, JSON |
Mistral OCR 25.03 | [Microsoft Managed Countries/Regions](/partner-center/marketplace/tax-details-marketplace#microsoft-managed-countriesregions) <br> Brazil <br> Hong Kong SAR <br> Israel | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | Not available |
97
97
Mistral Nemo | [Microsoft Managed Countries/Regions](/partner-center/marketplace/tax-details-marketplace#microsoft-managed-countriesregions) <br> Brazil <br> Hong Kong SAR <br> Israel | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | East US 2 <br> East US <br> North Central US <br> South Central US <br> West US <br> West US 3 |
98
98
Mistral Small 25.03 <br> Mistral Small | [Microsoft Managed Countries/Regions](/partner-center/marketplace/tax-details-marketplace#microsoft-managed-countriesregions) <br> Brazil <br> Hong Kong SAR <br> Israel | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | Not available |
99
+
Mistral Medium 3 (25.05) | [Microsoft Managed Countries/Regions](/partner-center/marketplace/tax-details-marketplace#microsoft-managed-countriesregions) <br> Brazil <br> Hong Kong SAR <br> Israel | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | Not available |
99
100
Ministral-3B | [Microsoft Managed Countries/Regions](/partner-center/marketplace/tax-details-marketplace#microsoft-managed-countriesregions) <br> Brazil <br> Hong Kong SAR<br> Israel | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | East US 2 <br> East US <br> North Central US <br> South Central US <br> West US <br> West US 3 |
100
101
Mistral Large (2407) <br> Mistral-Large | [Microsoft Managed Countries/Regions](/partner-center/marketplace/tax-details-marketplace#microsoft-managed-countriesregions) <br> Brazil <br> Hong Kong SAR<br> Israel | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | Not available |
101
102
Mistral-Large (2411) | [Microsoft Managed Countries/Regions](/partner-center/marketplace/tax-details-marketplace#microsoft-managed-countriesregions) <br> Brazil <br> Hong Kong SAR<br> Israel | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | East US 2 <br> East US <br> North Central US <br> South Central US <br> West US <br> West US 3 |
Copy file name to clipboardExpand all lines: articles/ai-services/content-understanding/concepts/standard-pro-modes.md
+5-1Lines changed: 5 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -82,10 +82,14 @@ Not sure which mode is right for your scenario? The following charts compare sta
82
82
83
83
You can try out the features of both Content Understanding standard and pro modes using the [Azure AI Foundry](https://ai.azure.com/explore/aiservices/vision/contentunderstanding). The service enables you to bring your own data and experiment with all the functionalities of both modes in a lightweight, no-code approach to help you find the best fit for your unique scenario.
84
84
85
-
### Pro mode known limitations
85
+
### Pro mode known limitations and best practices
86
86
87
87
* Content Understanding pro mode currently doesn't offer confidence scores or grounding. It currently supports generative and classification of your fields but doesn't support extraction only.
88
88
* Content Understanding pro mode is currently only available for documents.
89
+
* The system works in “lookup mode” on reference documents: don’t expect exhaustive information recovery; if this is desired, include the document in your input documents instead.
90
+
* Be as specific as possible with your schema (e.g., rather than identifying a list of inconsistencies, create a list field for each inconsistency type and describe it), and if possible, reference which parts of which documents should be consulted.
91
+
* Keep reference documents short and sweet: Limit documents to the critical ones, and keep them short where possible to improve recall.
Copy file name to clipboardExpand all lines: articles/ai-services/content-understanding/document/markdown.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -17,7 +17,7 @@ Azure AI Content Understanding converts unstructured documents into [GitHub Flav
17
17
18
18
## Words and selection marks
19
19
20
-
Recognized words and detected selection marks are represented in markdown as plain text. Content may be escaped to avoid ambiguity with markdown formatting syntax.
20
+
Recognized words and detected selection marks are represented in markdown as plain text. Selection marks are encoded using the Unicode characters `☒` (selected) and `☐` (unselected). Content might be escaped to avoid ambiguity with markdown formatting syntax.
21
21
22
22
## Barcodes
23
23
@@ -106,7 +106,7 @@ This is a footnote.
106
106
## Page metadata
107
107
108
108
Markdown doesn't natively encode page metadata, such as page numbers, headers, footers, and breaks.
109
-
Since this information may be useful for downstream applications, we encode such metadata as HTML comments.
109
+
Since this information might be useful for downstream applications, we encode such metadata as HTML comments.
Copy file name to clipboardExpand all lines: articles/ai-services/content-understanding/glossary.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -20,11 +20,11 @@ ms.author: lajanuar
20
20
|**Analyzer template**| A predefined configuration and field schema for an analyzer. It simplifies creating analyzers by allowing modifications to a template instead of starting from scratch. This feature is available only in [Azure AI Foundry portal](https://ai.azure.com/), not via REST API/SDKs. |
21
21
|**Analyzer result**| The output generated by an analyzer after processing input data. It typically includes extracted content in Markdown, extracted fields, and optional modality-specific details. |
22
22
|**Add-ons**| Added features that enhance content extraction results, such as layout elements, barcodes, and figures in documents. |
23
-
|**Fields**| List of structured key-value pairs derived from the content, as defined by the field schema. [Learn more about supported field value types.](service-limits.md)|
23
+
|**Fields**| List of structured key-value pairs derived from the content, as defined by the field schema. [Learn more about supported field value types.](service-limits.md#field-schema-limits)|
24
24
|**Field schema**| A formal description of the fields to extract from the input. It specifies the name, description, value type, generation method, and more for each field. |
25
25
|**Generation method**| The process of determining the extracted value of a specified field. Content Understanding supports: <br/> •**Extract**: Directly extract values from the input content, such as dates from receipts or item details from invoices. <br/> •**Classify**: Classify content into predefined categories, such as call sentiment or chart type. <br/> •**Generate**: Generate values from input data, such as summarizing an audio conversation or generating scene descriptions from videos. |
26
26
|**Span**| A reference indicating the location of an element (for example, field, word) within the extracted Markdown content. A character offset and length represent a span. Different programming languages use various character encodings, which can affect the exact offset and length values for Unicode text. To avoid confusion, spans are only returned if the desired encoding is explicitly specified in the request. Some elements can map to multiple spans if they aren't contiguous in the markdown (for example, page). |
27
-
| **Processing Location** | An API request parameter that defines the geographic region where Azure AI Services analyzes your data. You can choose from three options: `geography`, `dataZone`, and `global` to control where processing occurs. This setting helps meet data residency requirements and optimize performance or scalability based on your needs. For more information, *see* the Content Understanding API reference documentation.
27
+
| **Processing Location** | An API request parameter that defines the geographic region where Azure AI Services analyzes your data. You can choose from three options: `geography`, `dataZone`, and `global` to control where processing occurs. This setting helps meet data residency requirements and optimize performance or scalability based on your needs.
28
28
|**Grounding source**| The specific regions in content where a value was generated. It has different representations depending on the file type: <br>•**Image** - A polygon in the image, often an axis-aligned rectangle (bounding box). <br>•**PDF/TIFF** - A polygon on a specific page, often a quadrilateral. <br>•**Audio** - A start and end time range. <br>•**Video** - A start and end time range with an optional polygon in each frame, often a bounding box.|
29
29
|**Person directory**| A structured way to store face data for recognition tasks. You can add individual faces to the directory and later search for visually similar faces. You can also create person profiles, associate faces to them, and match new face images to known individuals. This setup supports both flexible face matching and identity recognition across images and videos. |
30
30
|**Confidence score**| The level of certainty that the extracted data is accurate. |
Copy file name to clipboardExpand all lines: articles/ai-services/content-understanding/image/overview.md
+2-4Lines changed: 2 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -42,12 +42,10 @@ Content Understanding offers several key benefits for extracting information fro
42
42
43
43
## Get started
44
44
45
-
Get started with processing images with Content Understanding by following our [REST API quickstart](LINK TO IMAGE TAB) or visiting [Azure AI Foundry](https://aka.ms/cu-landing) for a no code experience.
45
+
Get started with processing images with Content Understanding by following our [REST API quickstart](../quickstart/use-rest-api.md?tabs=image) or visiting [Azure AI Foundry](https://aka.ms/cu-landing) for a no code experience.
46
46
47
47
> [!NOTE]
48
-
> For best results, image schema should only be used to process non-document-based images.
49
-
> Text heavy images of documents should be processed using a document schema.
50
-
> Use cases that require extraction of text from document images or scanned documents should be processed using a document field extraction schema.
48
+
> Image analyzers are currently not optimized for scenarios where analysis is based primarily on extracted text. If your main goal is to extract and analyze text from images, consider using a document field extraction schema instead.
Copy file name to clipboardExpand all lines: articles/ai-services/content-understanding/whats-new.md
+12-9Lines changed: 12 additions & 9 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -14,22 +14,23 @@ ms.custom:
14
14
15
15
# What's new in Azure AI Content Understanding?
16
16
17
-
The Azure AI Content Understanding service is continuously updated. Bookmark this page to stay informed about the latest features and samples.
17
+
Azure AI Content Understanding service is updated on an ongoing basis. Bookmark this page to stay up to date with release notes, feature enhancements, and our newest documentation.
18
18
19
19
## May 2025
20
20
The Azure AI Content Understanding **`2025-05-01-preview`** REST API is now available. This update introduces the following updates and enhanced capabilities:
21
21
22
-
***Modes for documents**: With the **`2025-05-01-preview`** release, we introduce two modes: `standard` and `pro`. The `pro` mode, currently exclusive to the document analyzer, enables advanced capabilities. Content Understanding now supports reasoning across multiple documents as input for external knowledge, empowering users to derive agentic inferences directly from reference documents.
22
+
***Modes for documents**: With the **`2025-05-01-preview`** release, we introduce two modes: `standard` and `pro`. Content Understanding pro mode adds reasoning, support for multiple input documents, the ability to configure an external knowledge base for linking and validation. These features enable agentic processes for automating complex tasks requiring human effort. The `pro` mode is currently limited to the document analyzer, enables reasoning capabilities that can infer complicated outputs given multiple reference documents (for example, insurance policy documents) and multiple input documents (for example, forms and supporting documents that are case-specific)
23
23
24
24
***Document modality improvements** :
25
25
26
-
* Support for cross table extraction spanning multiple pages.
27
-
* Selection mark support for checkmark and radio buttons as unicode characters.
28
-
* Confidence score improvements.
29
-
* Support for embedded figures in Markdown as base64 string format.
26
+
* Improvements in content extraction:
27
+
* Support for table spanning multiple pages getting extracted as a single logical table.
28
+
* Selection mark support for checkmark and radio buttons as unicode characters.
29
+
* Bar code extraction as part of the default content extraction along with `OCR`.
30
+
* Support for embedded figures in Markdown as base64 string format output.
31
+
* Confidence score improvements with better grounding results for extractive fields.
30
32
* New file format support extended for following document types `.rtf`,`.txt`,`.xml`,`.json`, `.msg`,`.eml`,`.csv`, and `.xlsx`.
31
-
* Bar code extraction as part of the default content extraction along with `OCR`.
32
-
* Classification API for documents only. This API supports classifying and splitting documents with optional routing to analyzers.
33
+
* Classification API for documents only. This API supports classifying and splitting a single file containing multiple documents with optional routing to field extraction analyzers.
33
34
34
35
***Video Modality improvements**:
35
36
@@ -43,10 +44,12 @@ The Azure AI Content Understanding **`2025-05-01-preview`** REST API is now avai
43
44
***Face API preview**: Detection, recognition, and enrollment of faces.
44
45
45
46
***Billing Changes**: New simpler pricing model that lowers costs from the **2024-12-01-preview** REST API across all modalities.
47
+
48
+
***User experience improvements**: To improve latency issues, the UX experience is refreshed with ease-of-use and quick try-out experience as part of [Azure AI Foundry](https://aka.ms/cu-landing)
46
49
47
50
## April 2025
48
51
49
-
The Azure AI Content Understanding **2024-12-01-preview** REST API is now available. This update for preview API introduces the following updates and enhanced capabilities:
52
+
**2024-12-01-preview** REST API introduces the following updates and enhanced capabilities:
50
53
51
54
***General improvements**. For all modality, to request an increase from current limits, contact us at `[email protected]`.
52
55
***Prebuilt invoice template**. The invoice template is now customizable. Once you select the invoice template, you can access a predefined list of fields that can be tailored to your specific needs by adding or removing fields.
0 commit comments