Skip to content

Commit d7a64f3

Browse files
committed
Merge branch 'main' of https://github.com/microsoftdocs/azure-ai-docs-pr into mlflow-cli
2 parents dbbfb07 + d2c4967 commit d7a64f3

24 files changed

+387
-164
lines changed

articles/ai-services/document-intelligence/choose-model-feature.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ author: laujan
66
manager: nitinme
77
ms.service: azure-ai-document-intelligence
88
ms.topic: overview
9-
ms.date: 08/07/2024
9+
ms.date: 09/26/2024
1010
ms.author: lajanuar
1111
---
1212

articles/ai-services/document-intelligence/concept-analyze-document-response.md

Lines changed: 7 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -6,11 +6,10 @@ author: laujan
66
manager: nitinme
77
ms.service: azure-ai-document-intelligence
88
ms.topic: conceptual
9-
ms.date: 08/07/2024
9+
ms.date: 09/26/2024
1010
ms.author: vikurpad
1111
ms.custom:
1212
- references_regions
13-
- ignite-2023
1413
monikerRange: '>=doc-intel-3.0.0'
1514
---
1615

@@ -22,7 +21,7 @@ In this article, let's examine the different objects returned as part of the `An
2221

2322
## Analyze document request
2423

25-
The Document Intelligence APIs analyze images, PDFs, and other document files to extract and detect various content, layout, style, and semantic elements. The analyze operation is an async API. Submitting a document returns an **Operation-Location** header that contains the URL to poll for completion. When an analysis request completes successfully, the response contains the elements described in the [model data extraction](concept-model-overview.md#model-data-extraction).
24+
The Document Intelligence APIs analyze images, PDFs, and other document files to extract and detect various content, layout, style, and semantic elements. The `Analyze` operation is an async API. Submitting a document returns an **Operation-Location** header that contains the URL to poll for completion. When an analysis request completes successfully, the response contains the elements described in the [model data extraction](concept-model-overview.md#model-data-extraction).
2625

2726
### Response elements
2827

@@ -43,7 +42,7 @@ The top-level content property contains a concatenation of all content elements
4342

4443
## Analyze response
4544

46-
The analyze response for each API returns different objects. API responses contain elements from component models where applicable.
45+
The `Analyze` response for each API returns different objects. API responses contain elements from component models where applicable.
4746

4847
| Response content | Description | API |
4948
|--|--|--|
@@ -52,8 +51,8 @@ The analyze response for each API returns different objects. API responses conta
5251
| **styles**| Identified text element properties. | Read, Layout, General Document, Prebuilt, and Custom models|
5352
| **languages**| Identified language associated with each span of the text extracted | Read |
5453
| **tables**| Tabular content identified and extracted from the document. Tables relate to tables identified by the pretrained layout model. Content labeled as tables is extracted as structured fields in the documents object.| Layout, General Document, Invoice, and Custom models |
55-
| **figures**| Figures (charts, images) identified and extracted from the document, providing visual representations that aid in the understanding of complex information. | Layout model |
56-
| **sections** | Hierarchical document structure identified and extracted from the document. Section or subsection with the corresponding elements (paragraph, table, figure) attached to it. | Layout model |
54+
| **figures**| Figures (charts, images) identified and extracted from the document, providing visual representations that aid in the understanding of complex information. | The Layout model |
55+
| **sections** | Hierarchical document structure identified and extracted from the document. Section or subsection with the corresponding elements (paragraph, table, figure) attached to it. | The Layout model |
5756
| **keyValuePairs**| Key-value pairs recognized by a pretrained model. The key is a span of text from the document with the associated value. | General document and Invoice models |
5857
| **documents**| Fields recognized are returned in the `fields` dictionary within the list of documents| Prebuilt models, Custom models.|
5958

@@ -142,7 +141,7 @@ Based on its position and styling, a cell can be classified as general content,
142141
Figures (charts, images) in documents play a crucial role in complementing and enhancing the textual content, providing visual representations that aid in the understanding of complex information. The figures object detected by the Layout model has key properties like `boundingRegions` (the spatial locations of the figure on the document pages, including the page number and the polygon coordinates that outline the figure's boundary), `spans` (details the text spans related to the figure, specifying their offsets and lengths within the document's text. This connection helps in associating the figure with its relevant textual context), `elements` (the identifiers for text elements or paragraphs within the document that are related to or describe the figure) and `caption`, if any.
143142

144143
When *output=figures* is specified during the initial `Analyze` operation, the service generates cropped images for all detected figures that can be accessed via `/analyeResults/{resultId}/figures/{figureId}`.
145-
`FigureId` will be included in each figure object, following an undocumented convention of `{pageNumber}.{figureIndex}` where `figureIndex` resets to one per page.
144+
`FigureId` is included in each figure object, following an undocumented convention of `{pageNumber}.{figureIndex}` where `figureIndex` resets to one per page.
146145

147146
```json
148147
{
@@ -255,7 +254,7 @@ The semantic schema of a document type is described via the fields it contains.
255254
| date | Date | ISO 8601 - YYYY-MM-DD | InvoiceDate: "5/7/2022" → "2022-05-07" |
256255
| time | Time | ISO 8601 - hh:mm:ss | TransactionTime: "9:45 PM" → "21:45:00" |
257256
| phoneNumber | Phone number | E.164 - +{CountryCode}{SubscriberNumber} | WorkPhone: "(800) 555-7676" → "+18005557676"|
258-
| countryRegion | Country/region | ISO 3166-1 alpha-3 | CountryRegion: "United States" → "USA" |
257+
| countryRegion | Country/Region | ISO 3166-1 alpha-3 | CountryRegion: "United States" → "USA" |
259258
| selectionMark | Is selected | "signed" or "unsigned" | AcceptEula: ☑ → "selected" |
260259
| signature | Is signed | Same as content | LendeeSignature: {signature} → "signed" |
261260
| number | Floating point number | Floating point number | Quantity: "1.20" → 1.2|

articles/ai-services/document-intelligence/concept-custom-classifier.md

Lines changed: 5 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,16 +1,15 @@
11
---
2-
title: Custom classification model - Document Intelligence
2+
title: Custom classification model - Document Intelligence
33
titleSuffix: Azure AI services
44
description: Use the custom classification model to train a model to identify and split the documents you process within your application.
55
author: vkurpad
66
manager: nitinme
77
ms.service: azure-ai-document-intelligence
88
ms.topic: conceptual
9-
ms.date: 08/07/2024
9+
ms.date: 09/26/2024
1010
ms.author: lajanuar
1111
ms.custom:
1212
- references_regions
13-
- ignite-2023
1413
monikerRange: '>=doc-intel-3.1.0'
1514
---
1615

@@ -50,7 +49,6 @@ Custom classification models are deep-learning-model types that combine layout a
5049
Custom classification models can analyze a single- or multi-file documents to identify if any of the trained document types are contained within an input file. Here are the currently supported scenarios:
5150

5251
* A single file containing one document type, such as a loan application form.
53-
5452
* A single file containing multiple document types. For instance, a loan application package that contains a loan application form, payslip, and bank statement.
5553

5654
* A single file containing multiple instances of the same document. For instance, a collection of scanned invoices.
@@ -59,7 +57,8 @@ Custom classification models can analyze a single- or multi-file documents to id
5957

6058
✔️ The maximum allowed number of classes is `500`. The maximum allowed number of document samples per class is `100`.
6159

62-
The model classifies each page of the input document, unless specified, to one of the classes in the labeled dataset. You can specify the page numbers to analyze in the input document as well. To set the threshold for your application, use the confidence score from the response.
60+
The model classifies each page of the input document, unless specified, to one of the classes in the labeled dataset. You can specify the page numbers to analyze in the input document as well. To set the threshold for your application, use the confidence score from the response.
61+
6362
### Incremental training
6463

6564
With custom models, you need to maintain access to the training dataset to update your classifier with new samples for an existing class, or add new classes. Classifier models now support incremental training where you can reference an existing classifier and append new samples for an existing class or add new classes with samples. Incremental training enables scenarios where data retention is a challenge and the classifier needs to be updated to align with changing business needs. Incremental training is supported with models trained with API version `2024-02-29-preview` and later.
@@ -250,7 +249,7 @@ Alternatively, if you have a flat list of files or only plan to use a few select
250249
```
251250

252251
As an example, the file list `car-maint.jsonl` contains the following files.
253-
252+
254253
```json
255254
{"file":"classifier/car-maint/Commercial Motor Vehicle - Adatum.pdf"}
256255
{"file":"classifier/car-maint/Commercial Motor Vehicle - Fincher.pdf"}

articles/ai-services/document-intelligence/create-document-intelligence-resource.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ ms.service: azure-ai-document-intelligence
88
ms.custom:
99
- ignite-2023
1010
ms.topic: how-to
11-
ms.date: 04/24/2024
11+
ms.date: 09/26/2024
1212
ms.author: lajanuar
1313
---
1414

@@ -37,10 +37,10 @@ Let's get started:
3737

3838
## Create a resource
3939

40-
1. Next, you're going to fill out the **Create Document Intelligence** fields with the following values:
40+
1. Next, you're going to fill out the **`Create Document Intelligence`** fields with the following values:
4141

4242
* **Subscription**. Select your current subscription.
43-
* **Resource group**. The [Azure resource group](/azure/cloud-adoption-framework/govern/resource-consistency/resource-access-management#what-is-an-azure-resource-group) that contains your resource. You can create a new group or add it to a pre-existing group.
43+
* **Resource group**. The [Azure resource group](/azure/cloud-adoption-framework/govern/resource-consistency/resource-access-management#what-is-an-azure-resource-group) that contains your resource. You can create a new group or add it to an existing group.
4444
* **Region**. Select your local region.
4545
* **Name**. Enter a name for your resource. We recommend using a descriptive name, for example *YourNameFormRecognizer*.
4646
* **Pricing tier**. The cost of your resource depends on the pricing tier you choose and your usage. For more information, see [pricing details](https://azure.microsoft.com/pricing/details/cognitive-services/). You can use the free pricing tier (F0) to try the service, and upgrade later to a paid tier for production.

articles/ai-services/document-intelligence/quickstarts/includes/rest-api.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ ms.date: 09/09/2024
1010
ms.author: lajanuar
1111
---
1212

13-
<!-- markdownlint-disable MD036 -->
13+
<!-- markdownlint-disable MD036 -->
1414

1515
:::moniker range="doc-intel-4.0.0"
1616
| [Document Intelligence REST API](/rest/api/aiservices/operation-groups?view=rest-aiservices-v4.0%20(2024-07-31-preview)&preserve-view=true) | [Supported Azure SDKS](../../sdk-overview-v4-0.md)
@@ -109,7 +109,7 @@ Before you run the cURL command, make the following changes to the [post request
109109
:::moniker range="doc-intel-4.0.0"
110110

111111
```bash
112-
curl -v -i POST "{endpoint}/documentintelligence/documentModels/{modelId}:analyze?api-version=2024-02-29-preview" -H "Content-Type: application/json" -H "Ocp-Apim-Subscription-Key: {key}" --data-ascii "{'urlSource': '{your-document-url}'}"
112+
curl -v -i POST "{endpoint}/documentintelligence/documentModels/{modelId}:analyze?api-version=2024-07-31-preview" -H "Content-Type: application/json" -H "Ocp-Apim-Subscription-Key: {key}" --data-ascii "{'urlSource': '{your-document-url}'}"
113113
```
114114

115115
:::moniker-end
@@ -140,7 +140,7 @@ You receive a `202 (Success)` response that includes a read-only **Operation-Loc
140140

141141
:::moniker range="doc-intel-4.0.0"
142142

143-
After you call the [`Analyze document`](/rest/api/aiservices/document-models/analyze-document?view=rest-aiservices-2024-02-29-preview&preserve-view=true&branch=docintelligence&tabs=HTTP) API, call the [**Get analyze result**](/rest/api/aiservices/document-models/get-analyze-result?view=rest-aiservices-2024-02-29-preview&preserve-view=true&branch=docintelligence&tabs=HTTP) API to get the status of the operation and the extracted data. Before you run the command, make these changes:
143+
After you call the [`Analyze document`](/rest/api/aiservices/document-models/analyze-document?view=rest-aiservices-v4.0%20(2024-07-31-preview)&preserve-view=true&tabs=HTTP) API, call the [**Get analyze result**](/rest/api/aiservices/document-models/get-analyze-result?view=rest-aiservices-v4.0%20(2024-07-31-preview)&preserve-view=true&tabs=HTTP) API to get the status of the operation and the extracted data. Before you run the command, make these changes:
144144
:::moniker-end
145145

146146
:::moniker range="doc-intel-3.1.0"
@@ -165,7 +165,7 @@ After you call the [`Analyze document`](/rest/api/aiservices/document-models/ana
165165
:::moniker range="doc-intel-4.0.0"
166166

167167
```bash
168-
curl -v -X GET "{endpoint}/documentintelligence/documentModels/{modelId}/analyzeResults/{resultId}?api-version=2024-02-29-preview" -H "Ocp-Apim-Subscription-Key: {key}"
168+
curl -v -X GET "{endpoint}/documentintelligence/documentModels/{modelId}/analyzeResults/{resultId}?api-version=2024-07-31-preview" -H "Ocp-Apim-Subscription-Key: {key}"
169169
```
170170

171171
:::moniker-end
@@ -202,7 +202,7 @@ You receive a `200 (Success)` response with JSON output. The first field, `"stat
202202
"createdDateTime": "2024-03-25T19:31:37Z",
203203
"lastUpdatedDateTime": "2024-03-25T19:31:43Z",
204204
"analyzeResult": {
205-
"apiVersion": "2024-02-29-preview",
205+
"apiVersion": "2024-07-31-preview",
206206
"modelId": "prebuilt-invoice",
207207
"stringIndexType": "textElements"...
208208
..."pages": [

articles/ai-services/document-intelligence/service-limits.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
---
2-
title: Service quotas and limits - Document Intelligence
2+
title: Service quotas and limits - Document Intelligence
33
titleSuffix: Azure AI services
44
description: Quick reference, detailed description, and best practices for working within Azure AI Document Intelligence service Quotas and Limits
55
#services: cognitive-services
@@ -9,7 +9,7 @@ ms.service: azure-ai-document-intelligence
99
ms.custom:
1010
- ignite-2023
1111
ms.topic: conceptual
12-
ms.date: 06/26/2024
12+
ms.date: 09/26/2024
1313
ms.author: lajanuar
1414
monikerRange: '<=doc-intel-4.0.0'
1515
---

articles/ai-services/speech-service/speech-services-quotas-and-limits.md

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -136,6 +136,10 @@ The limits in this table apply per Speech resource when you create a personal vo
136136
| Quota | Free (F0)| Standard (S0) |
137137
|-----|-----|-----|
138138
| New connections per minute | Not available for F0 | 2 new connections per minute |
139+
| Max connection duration with speaking | Not available for F0 | 10 minutes<sup>1</sup> |
140+
| Max connection duration with idle state | Not available for F0 | 5 minutes |
141+
142+
<sup>1</sup> To ensure continuous operation of the real-time avatar for more than 10 minutes, you can enable auto-reconnect. For information about how to set up auto-reconnect, refer to this [sample code](https://github.com/Azure-Samples/cognitive-services-speech-sdk/blob/master/samples/js/browser/avatar/README.md) (search "auto reconnect").
139143

140144
#### Audio Content Creation tool
141145

articles/ai-services/speech-service/text-to-speech-avatar/real-time-synthesis-avatar.md

Lines changed: 14 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -180,6 +180,20 @@ avatarSynthesizer.speakTextAsync(spokenText).then(
180180

181181
You can find end-to-end working samples on [GitHub](https://github.com/Azure-Samples/cognitive-services-speech-sdk/tree/master/samples/js/browser/avatar).
182182

183+
## Close the real-time avatar connection
184+
185+
To avoid unnecessary costs after you finish using the real-time avatar, it’s important to close the connection. There are several ways to do this:
186+
187+
- When the browser web page is closed, the WebRTC client side peer connection object will be released, and the avatar connection will be automatically closed after a few seconds.
188+
- If the avatar remains idle for 5 minutes, the connection will be automatically closed by the avatar service.
189+
- You can proactively close the avatar connection by running the following code:
190+
191+
```javascript
192+
avatarSynthesizer.close()
193+
```
194+
195+
You can find end-to-end working samples on [GitHub](https://github.com/Azure-Samples/cognitive-services-speech-sdk/tree/master/samples/js/browser/avatar).
196+
183197
## Edit background
184198

185199
The avatar real-time synthesis API currently doesn't support setting a background image/video and only supports setting a solid-color background, without transparent background support. However, there's an alternative way to implement background customization on the client side, following these guidelines:

articles/ai-services/translator/custom-translator/overview.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ description: Custom Translator offers similar capabilities to what Microsoft Tra
55
author: laujan
66
manager: nitinme
77
ms.service: azure-ai-translator
8-
ms.date: 07/08/2024
8+
ms.date: 09/26/2024
99
ms.author: lajanuar
1010
ms.topic: overview
1111
---

articles/ai-services/translator/document-translation/reference/start-batch-translation.md

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ ms.author: lajanuar
88
author: laujan
99
ms.service: azure-ai-translator
1010
ms.topic: reference
11-
ms.date: 08/23/2024
11+
ms.date: 09/26/2024
1212
---
1313

1414
# Start batch translation
@@ -89,9 +89,6 @@ Request headers are:
8989
"storageType": "Folder"
9090
}
9191
],
92-
"options": {
93-
"experimental": true
94-
}
9592
}
9693

9794
```

0 commit comments

Comments
 (0)