Skip to content

Commit cfed402

Browse files
authored
Merge pull request #4475 from MicrosoftDocs/main
05/01/2025 PM Publishing
2 parents 3b62a39 + 09529f9 commit cfed402

File tree

8 files changed

+58
-50
lines changed

8 files changed

+58
-50
lines changed

articles/ai-services/content-understanding/concepts/analyzer-templates.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ ms.author: kabrow
77
manager: nitinme
88
ms.service: azure-ai-content-understanding
99
ms.topic: overview
10-
ms.date: 02/19/2025
10+
ms.date: 05/01/2025
1111
ms.custom: ignite-2024-understanding-release
1212
---
1313

@@ -26,7 +26,7 @@ The following analyzer templates are available for use in the [Azure AI Foundry
2626
|Template| Description|
2727
| ----|----|----|
2828
|Document analysis |Analyze documents to extract text, layout, structured fields, and more.|
29-
|Text analysis |Analyze texts and extract structured fields.|
29+
|Invoice analysis |Analyze invoice as prebuilt template and extract structured fields and tables.|
3030

3131
:::image type="content" source="../media/analyzer-template/scenario-template-document.png" alt-text="Screenshot of document analyzer template.":::
3232

articles/ai-services/content-understanding/document/overview.md

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ ms.author: lajanuar
77
manager: nitinme
88
ms.service: azure-ai-content-understanding
99
ms.topic: overview
10-
ms.date: 02/19/2025
10+
ms.date: 05/01/2025
1111
ms.custom: ignite-2024-understanding-release
1212
---
1313

@@ -23,6 +23,12 @@ Content Understanding is a cloud-based [Azure AI Service](../../what-are-ai-serv
2323

2424
Content Understanding enables organization to streamline data collection and processing, enhance operational efficiency, optimize data-driven decision making, and empower innovation. With customizable analyzers, Content Understanding allows for easy extraction of content or fields from documents and forms, tailored to specific business needs.
2525

26+
## April updates
27+
28+
* **Invoice prebuilt template**: Extract predefined schemas from various invoice formats. The out-of-the-box schema can be customized by adding or removing fields to suit your specific needs.
29+
30+
* **Generative and classify methods**: Added support for both generative and classification-based methods, enabling you to create generative fields such as summaries or categorize document details into multiple classes using the classify method.
31+
2632
## Business use cases
2733

2834
Document analyzers can process complex documents in various formats and templates:

articles/ai-services/content-understanding/service-limits.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -69,11 +69,11 @@ The following limits apply as of version 2024-12-01-preview.
6969

7070
| Property | Document | Image | Text | Audio | Video |
7171
| --- | --- | --- | --- | --- | --- |
72-
| Max fields | 50 | 10 | 10 | 10 | 10 |
72+
| Max fields | 50 | 50 | 50 | 50 | 50 |
7373
| Max classify field categories | 300 | 300 | 300 | 300 | 300 |
7474
| Supported generation methods | extract | generate<br>classify | generate<br>classify | generate<br>classify | generate<br>classify |
7575

76-
* The *Max fields* limit includes all named fields. For example, a list of strings counts as one field, while a group with string and number subfields counts as three fields. To extract beyond default limits, contact us at [email protected].
76+
* The *Max fields* limit includes all named fields. For example, a list of strings counts as one field, while a group with string and number subfields counts as three fields. To extend the limit for documents fields up to 100, contact us at `[email protected]`.
7777
* The *Max classify field categories* limit is the total number of categories across all fields using the `classify` generation method.
7878
* The generation method currently applies only to basic fields.
7979

@@ -83,7 +83,7 @@ The following limits apply as of version 2024-12-01-preview.
8383
| --- | --- | --- | --- | --- | --- |
8484
| Basic | No *boolean* | No *date*, *time* | *string* | *string* | No *date*, *time* |
8585
| List | N/A | No *date*, *time* | *string* | *string* | No *date*, *time* |
86-
| Group | N/A | No *date*, *time* | N/A | N/A | No *date*, *time* |
86+
| Group | N/A | No *date*, *time* |*string* | *string* | No *date*, *time* |
8787
| Table | No *boolean* | No *date*, *time* | *string* | *string* | No *date*, *time* |
8888
| Fixed table | No *boolean* | N/A | N/A | N/A | N/A |
8989

articles/ai-services/content-understanding/whats-new.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -41,4 +41,5 @@ The Content Understanding **2024-12-01-preview** REST API is now available. This
4141
* Add download code samples for quick setup added.
4242

4343
## November 2024
44-
Welcome! The Azure AI Content Understanding API version `2024-12-01-preview` is now in public preview. This version allows you to generate a structured representation of content tailored to specific tasks from various modalities or formats. Content Understanding uses a defined schema to extract content suitable for processing by large language models and subsequent applications.
44+
45+
Welcome! The Azure AI Content Understanding API version `2024-12-01-preview` is now in public preview. This version allows you to generate a structured representation of content tailored to specific tasks from various modalities or formats. Content Understanding uses a defined schema to extract content suitable for processing by large language models and subsequent applications.

articles/ai-services/openai/concepts/models.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@ Azure OpenAI Service is powered by a diverse set of models with different capabi
4343

4444
| Model ID | Description | Context Window | Max Output Tokens | Training Data (up to) |
4545
| --- | :--- |:--- |:---|:---: |
46-
| `gpt-4.1` (2025-04-14) | - Text & image input <br> - Text output <br> - Chat completions API <br>- Responses API <br> - Streaming <br> - Function calling <br> Structured outputs (chat completions) | 1,047,576 | 32,768 | May 31, 2024 |
46+
| `gpt-4.1` (2025-04-14) | - Text & image input <br> - Text output <br> - Chat completions API <br>- Responses API <br> - Streaming <br> - Function calling <br> Structured outputs (chat completions) | - 1,047,576 <br> - 128,000 (provisioned managed deployments) | 32,768 | May 31, 2024 |
4747
| `gpt-4.1-nano` (2025-04-14) <br><br> **Fastest 4.1 model** | - Text & image input <br> - Text output <br> - Chat completions API <br>- Responses API <br> - Streaming <br> - Function calling <br> Structured outputs (chat completions) | 1,047,576 | 32,768 | May 31, 2024 |
4848
| `gpt-4.1-mini` (2025-04-14) | - Text & image input <br> - Text output <br> - Chat completions API <br>- Responses API <br> - Streaming <br> - Function calling <br> Structured outputs (chat completions) | 1,047,576 | 32,768 | May 31, 2024 |
4949

articles/ai-services/openai/concepts/provisioned-throughput.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ title: Azure OpenAI Service provisioned throughput
33
description: Learn about provisioned throughput and Azure OpenAI.
44
ms.service: azure-ai-openai
55
ms.topic: conceptual
6-
ms.date: 03/31/2025
6+
ms.date: 04/30/2025
77
manager: nitinme
88
author: aahill #ChrisHMSFT
99
ms.author: aahi #chrhoder

articles/ai-services/openai/how-to/provisioned-throughput-onboarding.md

Lines changed: 13 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -72,23 +72,24 @@ Customers that require long-term usage of provisioned, data zoned provisioned, a
7272
> Charges for deployments on a deleted resource will continue until the resource is purged. To prevent this, delete a resource’s deployment before deleting the resource. For more information, see [Recover or purge deleted Azure AI services resources](../../recover-purge-resources.md).
7373
7474
## How much throughput per PTU you get for each model
75-
The amount of throughput (measured in tokens per minute or TPM) a deployment gets per PTU is a function of the input and output tokens in a given minute.
7675

77-
Generating output tokens requires more processing than input tokens. For the models specified in the table below, 1 output token counts as 3 input tokens towards your TPM-per-PTU limit. The service dynamically balances the input & output costs, so users do not have to set specific input and output limits. This approach means your deployment is resilient to fluctuations in the workload.
7876

79-
To help with simplifying the sizing effort, the following table outlines the TPM-per-PTU for the specified models. To understand the impact of output tokens on the TPM-per-PTU limit, use the 3 input token to 1 output token ratio.
8077

81-
For a detailed understanding of how different ratios of input and output tokens impact the throughput your workload needs, see the [Azure OpenAI capacity calculator](https://ai.azure.com/resource/calculator). The table also shows Service Level Agreement (SLA) Latency Target Values per model. For more information about the SLA for Azure OpenAI Service, see the [Service Level Agreements (SLA) for Online Services page](https://www.microsoft.com/licensing/docs/view/Service-Level-Agreements-SLA-for-Online-Services?lang=1)
8278

8379

84-
|Topic| **gpt-4o** | **gpt-4o-mini** | **o1**|
85-
| --- | --- | --- | --- |
86-
|Global & data zone provisioned minimum deployment|15|15|15|
87-
|Global & data zone provisioned scale increment|5|5|5|
88-
|Regional provisioned minimum deployment|50|25|50|
89-
|Regional provisioned scale increment|50|25|50|
90-
|Input TPM per PTU |2,500|37,000|230|
91-
|Latency Target Value |25 Tokens Per Second|33 Tokens Per Second|25 Tokens Per Second|
80+
The amount of throughput (measured in tokens per minute or TPM) a deployment gets per PTU is a function of the input and output tokens in a given minute. Generating output tokens requires more processing than input tokens.  Starting with GPT 4.1 models and later, the system matches the global standard price ratio between input and output tokens. Cached tokens are deducted 100% from the utilization.
81+
82+
For example, for `gpt-4.1:2025-04-14`, 1 output token counts as 4 input tokens towards your utilization limit which matches the [pricing](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service/). Older models use a different ratio and for a deeper understanding on how different ratios of input and output tokens impact the throughput your workload needs, see the [Azure OpenAI capacity calculator](https://ai.azure.com/resource/calculator).
83+
84+
85+
|Topic| **gpt-4o** | **gpt-4o-mini** | **o1**| gpt-4.1 |
86+
| --- | --- | --- | --- | --- |
87+
|Global & data zone provisioned minimum deployment|15|15|15|15 |
88+
|Global & data zone provisioned scale increment|5|5|5| 5 |
89+
|Regional provisioned minimum deployment|50|25|50| 50 |
90+
|Regional provisioned scale increment|50|25|50| 50 |
91+
|Input TPM per PTU |2,500|37,000|230| 3000 |
92+
|Latency Target Value |25 Tokens Per Second|33 Tokens Per Second|25 Tokens Per Second| 44 Tokens Per Second |
9293

9394
For a full list, see the [Azure OpenAI Service in Azure AI Foundry portal calculator](https://ai.azure.com/resource/calculator).
9495

0 commit comments

Comments
 (0)