Skip to content

Commit 67c9744

Browse files
committed
Merge branch 'patch-14' into kate-4954-4949-overview-overview
2 parents 60b617a + 53218ab commit 67c9744

File tree

9 files changed

+29
-21
lines changed

9 files changed

+29
-21
lines changed

articles/ai-foundry/concepts/models-featured.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -250,7 +250,7 @@ See [the Microsoft model collection in Azure AI Foundry portal](https://ai.azure
250250

251251
Mistral AI offers two categories of models, namely:
252252

253-
- _Premium models_: These include Mistral Large, Mistral Small, Mistral-OCR-2503, and Ministral 3B models, and are available as serverless APIs with pay-as-you-go token-based billing.
253+
- _Premium models_: These include Mistral Large, Mistral Small, Mistral-OCR-2503, Mistral Medium 3 (25.05), and Ministral 3B models, and are available as serverless APIs with pay-as-you-go token-based billing.
254254
- _Open models_: These include Mistral-small-2503, Codestral, and Mistral Nemo (that are available as serverless APIs with pay-as-you-go token-based billing), and [Mixtral-8x7B-Instruct-v01, Mixtral-8x7B-v01, Mistral-7B-Instruct-v01, and Mistral-7B-v01](../how-to/deploy-models-mistral-open.md)(that are available to download and run on self-hosted managed endpoints).
255255

256256

@@ -262,6 +262,7 @@ Mistral AI offers two categories of models, namely:
262262
| [Mistral-Large-2411](https://ai.azure.com/explore/models/Mistral-Large-2411/version/2/registry/azureml-mistral) | [chat-completion](../model-inference/how-to/use-chat-completions.md?context=/azure/ai-foundry/context/context) | - **Input:** text (128,000 tokens) <br /> - **Output:** text (4,096 tokens) <br /> - **Tool calling:** Yes <br /> - **Response formats:** Text, JSON |
263263
| [Mistral-large-2407](https://ai.azure.com/explore/models/Mistral-large-2407/version/1/registry/azureml-mistral) <br /> (deprecated) | [chat-completion](../model-inference/how-to/use-chat-completions.md?context=/azure/ai-foundry/context/context) | - **Input:** text (131,072 tokens) <br /> - **Output:** text (4,096 tokens) <br /> - **Tool calling:** Yes <br /> - **Response formats:** Text, JSON |
264264
| [Mistral-large](https://ai.azure.com/explore/models/Mistral-large/version/1/registry/azureml-mistral) <br /> (deprecated) | [chat-completion](../model-inference/how-to/use-chat-completions.md?context=/azure/ai-foundry/context/context) | - **Input:** text (32,768 tokens) <br /> - **Output:** text (4,096 tokens) <br /> - **Tool calling:** Yes <br /> - **Response formats:** Text, JSON |
265+
| [Mistral-medium-2505](https://aka.ms/aistudio/landing/mistral-medium-2505) | [chat-completion](../model-inference/how-to/use-chat-completions.md?context=/azure/ai-foundry/context/context) | - **Input:** text (128,000 tokens), image <br /> - **Output:** text (128,000 tokens) <br /> - **Tool calling:** No <br /> - **Response formats:** Text, JSON |
265266
| [Mistral-OCR-2503](https://aka.ms/aistudio/landing/mistral-ocr-2503) | [image to text](../how-to/use-image-models.md) | - **Input:** image or PDF pages (1,000 pages, max 50MB PDF file) <br> - **Output:** text <br /> - **Tool calling:** No <br /> - **Response formats:** Text, JSON, Markdown |
266267
| [Mistral-small-2503](https://aka.ms/aistudio/landing/mistral-small-2503) | [chat-completion (with images)](../model-inference/how-to/use-chat-multi-modal.md?context=/azure/ai-foundry/context/context) | - **Input:** text and images (131,072 tokens), <br> image-based tokens are 16px x 16px <br> blocks of the original images <br /> - **Output:** text (4,096 tokens) <br /> - **Tool calling:** Yes <br /> - **Response formats:** Text, JSON |
267268
| [Mistral-small](https://ai.azure.com/explore/models/Mistral-small/version/1/registry/azureml-mistral) | [chat-completion](../model-inference/how-to/use-chat-completions.md?context=/azure/ai-foundry/context/context) | - **Input:** text (32,768 tokens) <br /> - **Output:** text (4,096 tokens) <br /> - **Tool calling:** Yes <br /> - **Response formats:** Text, JSON |

articles/ai-foundry/includes/region-availability-maas.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -96,6 +96,7 @@ Codestral-2501 | [Microsoft Managed Countries/Regions](/partner-center/market
9696
Mistral OCR 25.03 | [Microsoft Managed Countries/Regions](/partner-center/marketplace/tax-details-marketplace#microsoft-managed-countriesregions) <br> Brazil <br> Hong Kong SAR <br> Israel | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | Not available |
9797
Mistral Nemo | [Microsoft Managed Countries/Regions](/partner-center/marketplace/tax-details-marketplace#microsoft-managed-countriesregions) <br> Brazil <br> Hong Kong SAR <br> Israel | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | East US 2 <br> East US <br> North Central US <br> South Central US <br> West US <br> West US 3 |
9898
Mistral Small 25.03 <br> Mistral Small | [Microsoft Managed Countries/Regions](/partner-center/marketplace/tax-details-marketplace#microsoft-managed-countriesregions) <br> Brazil <br> Hong Kong SAR <br> Israel | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | Not available |
99+
Mistral Medium 3 (25.05) | [Microsoft Managed Countries/Regions](/partner-center/marketplace/tax-details-marketplace#microsoft-managed-countriesregions) <br> Brazil <br> Hong Kong SAR <br> Israel | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | Not available |
99100
Ministral-3B | [Microsoft Managed Countries/Regions](/partner-center/marketplace/tax-details-marketplace#microsoft-managed-countriesregions) <br> Brazil <br> Hong Kong SAR<br> Israel | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | East US 2 <br> East US <br> North Central US <br> South Central US <br> West US <br> West US 3 |
100101
Mistral Large (2407) <br> Mistral-Large | [Microsoft Managed Countries/Regions](/partner-center/marketplace/tax-details-marketplace#microsoft-managed-countriesregions) <br> Brazil <br> Hong Kong SAR<br> Israel | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | Not available |
101102
Mistral-Large (2411) | [Microsoft Managed Countries/Regions](/partner-center/marketplace/tax-details-marketplace#microsoft-managed-countriesregions) <br> Brazil <br> Hong Kong SAR<br> Israel | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | East US 2 <br> East US <br> North Central US <br> South Central US <br> West US <br> West US 3 |

articles/ai-services/content-understanding/concepts/standard-pro-modes.md

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -82,10 +82,14 @@ Not sure which mode is right for your scenario? The following charts compare sta
8282

8383
You can try out the features of both Content Understanding standard and pro modes using the [Azure AI Foundry](https://ai.azure.com/explore/aiservices/vision/contentunderstanding). The service enables you to bring your own data and experiment with all the functionalities of both modes in a lightweight, no-code approach to help you find the best fit for your unique scenario.
8484

85-
### Pro mode known limitations
85+
### Pro mode known limitations and best practices
8686

8787
* Content Understanding pro mode currently doesn't offer confidence scores or grounding. It currently supports generative and classification of your fields but doesn't support extraction only.
8888
* Content Understanding pro mode is currently only available for documents.
89+
* The system works in “lookup mode” on reference documents: don’t expect exhaustive information recovery; if this is desired, include the document in your input documents instead.
90+
* Be as specific as possible with your schema (e.g., rather than identifying a list of inconsistencies, create a list field for each inconsistency type and describe it), and if possible, reference which parts of which documents should be consulted.
91+
* Keep reference documents short and sweet: Limit documents to the critical ones, and keep them short where possible to improve recall.
92+
8993

9094
## Next steps
9195

articles/ai-services/content-understanding/document/markdown.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ Azure AI Content Understanding converts unstructured documents into [GitHub Flav
1717

1818
## Words and selection marks
1919

20-
Recognized words and detected selection marks are represented in markdown as plain text. Content may be escaped to avoid ambiguity with markdown formatting syntax.
20+
Recognized words and detected selection marks are represented in markdown as plain text. Selection marks are encoded using the Unicode characters `` (selected) and `` (unselected). Content might be escaped to avoid ambiguity with markdown formatting syntax.
2121

2222
## Barcodes
2323

@@ -106,7 +106,7 @@ This is a footnote.
106106
## Page metadata
107107

108108
Markdown doesn't natively encode page metadata, such as page numbers, headers, footers, and breaks.
109-
Since this information may be useful for downstream applications, we encode such metadata as HTML comments.
109+
Since this information might be useful for downstream applications, we encode such metadata as HTML comments.
110110

111111
| Metadata | Markdown |
112112
| --- | --- |

articles/ai-services/content-understanding/document/overview.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -98,4 +98,4 @@ Developers using Content Understanding should review Microsoft's policies on cus
9898
* Try processing your document content using Content Understanding in [Azure AI Foundry](https://aka.ms/cu-landing).
9999
* Learn to analyze document content [**analyzer templates**](../quickstart/use-ai-foundry.md).
100100
* Review code samples: [**visual document search**](https://github.com/Azure-Samples/azure-ai-search-with-content-understanding-python/blob/main/notebooks/search_with_visual_document.ipynb).
101-
* Review code sample: [**analyzer templates**](https://github.com/Azure-Samples/azure-ai-content-understanding-python/tree/main/analyzer_templates).
101+

articles/ai-services/content-understanding/glossary.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -20,11 +20,11 @@ ms.author: lajanuar
2020
| **Analyzer template** | A predefined configuration and field schema for an analyzer. It simplifies creating analyzers by allowing modifications to a template instead of starting from scratch. This feature is available only in [Azure AI Foundry portal](https://ai.azure.com/), not via REST API/SDKs. |
2121
| **Analyzer result** | The output generated by an analyzer after processing input data. It typically includes extracted content in Markdown, extracted fields, and optional modality-specific details. |
2222
| **Add-ons** | Added features that enhance content extraction results, such as layout elements, barcodes, and figures in documents. |
23-
| **Fields** | List of structured key-value pairs derived from the content, as defined by the field schema. [Learn more about supported field value types.](service-limits.md) |
23+
| **Fields** | List of structured key-value pairs derived from the content, as defined by the field schema. [Learn more about supported field value types.](service-limits.md#field-schema-limits) |
2424
| **Field schema** | A formal description of the fields to extract from the input. It specifies the name, description, value type, generation method, and more for each field. |
2525
| **Generation method** | The process of determining the extracted value of a specified field. Content Understanding supports: <br/> &bullet; **Extract**: Directly extract values from the input content, such as dates from receipts or item details from invoices. <br/> &bullet; **Classify**: Classify content into predefined categories, such as call sentiment or chart type. <br/> &bullet; **Generate**: Generate values from input data, such as summarizing an audio conversation or generating scene descriptions from videos. |
2626
| **Span** | A reference indicating the location of an element (for example, field, word) within the extracted Markdown content. A character offset and length represent a span. Different programming languages use various character encodings, which can affect the exact offset and length values for Unicode text. To avoid confusion, spans are only returned if the desired encoding is explicitly specified in the request. Some elements can map to multiple spans if they aren't contiguous in the markdown (for example, page). |
27-
| **Processing Location** | An API request parameter that defines the geographic region where Azure AI Services analyzes your data. You can choose from three options: `geography`, `dataZone`, and `global` to control where processing occurs. This setting helps meet data residency requirements and optimize performance or scalability based on your needs. For more information, *see* the Content Understanding API reference documentation.
27+
| **Processing Location** | An API request parameter that defines the geographic region where Azure AI Services analyzes your data. You can choose from three options: `geography`, `dataZone`, and `global` to control where processing occurs. This setting helps meet data residency requirements and optimize performance or scalability based on your needs.
2828
| **Grounding source** | The specific regions in content where a value was generated. It has different representations depending on the file type: <br>&bullet; **Image** - A polygon in the image, often an axis-aligned rectangle (bounding box). <br>&bullet; **PDF/TIFF** - A polygon on a specific page, often a quadrilateral. <br>&bullet; **Audio** - A start and end time range. <br>&bullet; **Video** - A start and end time range with an optional polygon in each frame, often a bounding box.|
2929
| **Person directory** | A structured way to store face data for recognition tasks. You can add individual faces to the directory and later search for visually similar faces. You can also create person profiles, associate faces to them, and match new face images to known individuals. This setup supports both flexible face matching and identity recognition across images and videos. |
3030
| **Confidence score** | The level of certainty that the extracted data is accurate. |

articles/ai-services/content-understanding/image/overview.md

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -42,12 +42,10 @@ Content Understanding offers several key benefits for extracting information fro
4242

4343
## Get started
4444

45-
Get started with processing images with Content Understanding by following our [REST API quickstart](LINK TO IMAGE TAB) or visiting [Azure AI Foundry](https://aka.ms/cu-landing) for a no code experience.
45+
Get started with processing images with Content Understanding by following our [REST API quickstart](../quickstart/use-rest-api.md?tabs=image) or visiting [Azure AI Foundry](https://aka.ms/cu-landing) for a no code experience.
4646

4747
> [!NOTE]
48-
> For best results, image schema should only be used to process non-document-based images.
49-
> Text heavy images of documents should be processed using a document schema.
50-
> Use cases that require extraction of text from document images or scanned documents should be processed using a document field extraction schema.
48+
> Image analyzers are currently not optimized for scenarios where analysis is based primarily on extracted text. If your main goal is to extract and analyze text from images, consider using a document field extraction schema instead.
5149
5250

5351
> [!IMPORTANT]

articles/ai-services/content-understanding/whats-new.md

Lines changed: 12 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -14,22 +14,23 @@ ms.custom:
1414

1515
# What's new in Azure AI Content Understanding?
1616

17-
The Azure AI Content Understanding service is continuously updated. Bookmark this page to stay informed about the latest features and samples.
17+
Azure AI Content Understanding service is updated on an ongoing basis. Bookmark this page to stay up to date with release notes, feature enhancements, and our newest documentation.
1818

1919
## May 2025
2020
The Azure AI Content Understanding **`2025-05-01-preview`** REST API is now available. This update introduces the following updates and enhanced capabilities:
2121

22-
* **Modes for documents**: With the **`2025-05-01-preview`** release, we introduce two modes: `standard` and `pro`. The `pro` mode, currently exclusive to the document analyzer, enables advanced capabilities. Content Understanding now supports reasoning across multiple documents as input for external knowledge, empowering users to derive agentic inferences directly from reference documents.
22+
* **Modes for documents**: With the **`2025-05-01-preview`** release, we introduce two modes: `standard` and `pro`. Content Understanding pro mode adds reasoning, support for multiple input documents, the ability to configure an external knowledge base for linking and validation. These features enable agentic processes for automating complex tasks requiring human effort. The `pro` mode is currently limited to the document analyzer, enables reasoning capabilities that can infer complicated outputs given multiple reference documents (for example, insurance policy documents) and multiple input documents (for example, forms and supporting documents that are case-specific)
2323

2424
* **Document modality improvements** :
2525

26-
* Support for cross table extraction spanning multiple pages.
27-
* Selection mark support for checkmark and radio buttons as unicode characters.
28-
* Confidence score improvements.
29-
* Support for embedded figures in Markdown as base64 string format.
26+
* Improvements in content extraction:
27+
* Support for table spanning multiple pages getting extracted as a single logical table.
28+
* Selection mark support for checkmark and radio buttons as unicode characters.
29+
* Bar code extraction as part of the default content extraction along with `OCR`.
30+
* Support for embedded figures in Markdown as base64 string format output.
31+
* Confidence score improvements with better grounding results for extractive fields.
3032
* New file format support extended for following document types `.rtf`,`.txt`,`.xml`,`.json`, `.msg`,`.eml`,`.csv`, and `.xlsx`.
31-
* Bar code extraction as part of the default content extraction along with `OCR`.
32-
* Classification API for documents only. This API supports classifying and splitting documents with optional routing to analyzers.
33+
* Classification API for documents only. This API supports classifying and splitting a single file containing multiple documents  with optional routing to field extraction analyzers.
3334

3435
* **Video Modality improvements**:
3536

@@ -43,10 +44,12 @@ The Azure AI Content Understanding **`2025-05-01-preview`** REST API is now avai
4344
* **Face API preview**: Detection, recognition, and enrollment of faces.
4445

4546
* **Billing Changes**: New simpler pricing model that lowers costs from the **2024-12-01-preview** REST API across all modalities.
47+
48+
* **User experience improvements**: To improve latency issues, the UX experience is refreshed with ease-of-use and quick try-out experience as part of [Azure AI Foundry](https://aka.ms/cu-landing)
4649

4750
## April 2025
4851

49-
The Azure AI Content Understanding **2024-12-01-preview** REST API is now available. This update for preview API introduces the following updates and enhanced capabilities:
52+
**2024-12-01-preview** REST API introduces the following updates and enhanced capabilities:
5053

5154
* **General improvements**. For all modality, to request an increase from current limits, contact us at `[email protected]`.
5255
* **Prebuilt invoice template**. The invoice template is now customizable. Once you select the invoice template, you can access a predefined list of fields that can be tailored to your specific needs by adding or removing fields.

0 commit comments

Comments
 (0)