Skip to content

Commit ff6c771

Browse files
authored
Merge branch 'MicrosoftDocs:main' into heidist-free
2 parents 62b56f7 + c3664e0 commit ff6c771

File tree

283 files changed

+3547
-1704
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

283 files changed

+3547
-1704
lines changed

articles/ai-services/.openpublishing.redirection.ai-services.json

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -400,6 +400,11 @@
400400
"redirect_url": "/azure/ai-services/speech-service/quickstarts/setup-platform",
401401
"redirect_document_id": false
402402
},
403+
{
404+
"source_path_from_root": "/articles/ai-services/speech-service/devices-sdk-release-notes.md",
405+
"redirect_url": "/azure/ai-services/speech-service/release-notes",
406+
"redirect_document_id": false
407+
},
403408
{
404409
"source_path_from_root": "/articles/ai-services/anomaly-detector/how-to/postman.md",
405410
"redirect_url": "/azure/ai-services/anomaly-detector/overview",

articles/ai-services/custom-vision-service/overview.md

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -18,8 +18,7 @@ keywords: image recognition, image identifier, image recognition app, custom vis
1818

1919
Azure AI Custom Vision is an image recognition service that lets you build, deploy, and improve your own **image identifier** models. An image identifier applies labels to images, according to their visual characteristics. Each label represents a classification or object. Custom Vision allows you to specify your own labels and train custom models to detect them.
2020

21-
> [!TIP]
22-
> The Azure AI vision Image Analysis API, based on Florence foundational model, now supports custom models with few-shot learning capability. [Use Image Analysis 4.0](../computer-vision/how-to/model-customization.md) to create custom image identifier models using the latest technology from Azure. To migrate a Custom Vision project to the new Image Analysis 4.0 system, see the [Migration guide](../computer-vision/how-to/migrate-from-custom-vision.md). To compare the two services, see the [Comparison page](./concepts/compare-alternatives.md).
21+
2322

2423

2524
You can use Custom Vision through a client library SDK, REST API, or through the [Custom Vision web portal](https://customvision.ai/). Follow a quickstart to get started.

articles/ai-services/language-service/conversational-language-understanding/concepts/best-practices.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -47,7 +47,7 @@ You also want to avoid mixing different schema designs. Don't build half of your
4747

4848
## Use standard training before advanced training
4949

50-
[Standard training](../how-to/train-model.md#training-modes) is free and faster than advanced training. It can help you quickly understand the effect of changing your training set or schema while you build the model. After you're satisfied with the schema, consider using advanced training to get the best AIQ out of your model.
50+
[Standard training](../how-to/train-model.md#training-modes) is free and faster than advanced training. It can help you quickly understand the effect of changing your training set or schema while you build the model. After you're satisfied with the schema, consider using advanced training to get the best model quality.
5151

5252
## Use the evaluation feature
5353

@@ -113,7 +113,7 @@ If you enable this feature, the utterance count of your training set increases.
113113

114114
## Address model overconfidence
115115

116-
Customers can use the LoraNorm recipe version if the model is being incorrectly overconfident. An example of this behavior can be like the following scenario where the model predicts the incorrect intent with 100% confidence. This score makes the confidence threshold project setting unusable.
116+
Customers can use the LoraNorm traning configuration version if the model is being incorrectly overconfident. An example of this behavior can be like the following scenario where the model predicts the incorrect intent with 100% confidence. This score makes the confidence threshold project setting unusable.
117117

118118
| Text | Predicted intent | Confidence score |
119119
|----|----|----|
@@ -243,7 +243,7 @@ curl --request POST \
243243

244244
## Address out-of-domain utterances
245245

246-
Customers can use the newly updated recipe version `2024-08-01-preview` (previously `2024-06-01-preview`) if the model has poor AIQ on out-of-domain utterances. An example of this scenario with the default recipe can be like the following example where the model has three intents: `Sports`, `QueryWeather`, and `Alarm`. The test utterances are out-of-domain utterances and the model classifies them as `InDomain` with a relatively high confidence score.
246+
Customers can use the newly updated training configuration version `2024-08-01-preview` (previously `2024-06-01-preview`) if the model has poor quality on out-of-domain utterances. An example of this scenario with the default training configuration can be like the following example where the model has three intents: `Sports`, `QueryWeather`, and `Alarm`. The test utterances are out-of-domain utterances and the model classifies them as `InDomain` with a relatively high confidence score.
247247

248248
| Text | Predicted intent | Confidence score |
249249
|----|----|----|
@@ -273,6 +273,6 @@ After the request is sent, you can track the progress of the training job in Lan
273273

274274
Caveats:
275275

276-
- The None score threshold for the app (confidence threshold below which `topIntent` is marked as `None`) when you use this recipe should be set to 0. This setting is used because this new recipe attributes a certain portion of the in-domain probabilities to out of domain so that the model isn't incorrectly overconfident about in-domain utterances. As a result, users might see slightly reduced confidence scores for in-domain utterances as compared to the prod recipe.
277-
- We don't recommend this recipe for apps with only two intents, such as `IntentA` and `None`, for example.
278-
- We don't recommend this recipe for apps with a low number of utterances per intent. We highly recommend a minimum of 25 utterances per intent.
276+
- The None score threshold for the app (confidence threshold below which `topIntent` is marked as `None`) when you use this training configuration should be set to 0. This setting is used because this new training configuration attributes a certain portion of the in-domain probabilities to out of domain so that the model isn't incorrectly overconfident about in-domain utterances. As a result, users might see slightly reduced confidence scores for in-domain utterances as compared to the prod training configuration.
277+
- We don't recommend this training configuration for apps with only two intents, such as `IntentA` and `None`, for example.
278+
- We don't recommend this training configuration for apps with a low number of utterances per intent. We highly recommend a minimum of 25 utterances per intent.

articles/ai-services/language-service/summarization/how-to/document-summarization.md

Lines changed: 82 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -122,22 +122,11 @@ curl -i -X POST https://<your-language-resource-endpoint>/language/analyze-text/
122122
{
123123
"kind": "AbstractiveSummarization",
124124
"taskName": "Text Abstractive Summarization Task 1",
125-
"parameters": {
126-
"summaryLength": "short"
127-
}
128125
}
129126
]
130127
}
131128
'
132129
```
133-
If you don't specify `summaryLength`, the model determines the summary length.
134-
135-
### Using the summaryLength parameter
136-
For the `summaryLength` parameter, three values are accepted:
137-
* oneSentence: Generates a summary of mostly 1 sentence, with around 80 tokens.
138-
* short: Generates a summary of mostly 2-3 sentences, with around 120 tokens.
139-
* medium: Generates a summary of mostly 4-6 sentences, with around 170 tokens.
140-
* long: Generates a summary of mostly over 7 sentences, with around 210 tokens.
141130

142131
2. Make the following changes in the command where needed:
143132
- Replace the value `your-language-resource-key` with your key.
@@ -222,7 +211,14 @@ The following cURL commands are executed from a BASH shell. Edit these commands
222211

223212
The query-based text summarization API is an extension to the existing text summarization API.
224213

225-
The biggest difference is a new `query` field in the request body (under `tasks` > `parameters` > `query`). Additionally, there's a new way to specify the preferred `summaryLength` in "buckets" of short/medium/long, which we recommend using instead of `sentenceCount`, especially when using abstractive. Below is an example request:
214+
The biggest difference is a new `query` field in the request body (under `tasks` > `parameters` > `query`).
215+
216+
> [!TIP]
217+
> Query based summarization has some differentiation in the utilization of length control based on the type of query based summarization you're using:
218+
> - Query based extractive summarization supports length control by specifying sentenceCount.
219+
> - Query based abstractive summarization doesn't support length control.
220+
221+
Below is an example request:
226222

227223
```bash
228224
curl -i -X POST https://<your-language-resource-endpoint>/language/analyze-text/jobs?api-version=2023-11-15-preview \
@@ -253,22 +249,93 @@ curl -i -X POST https://<your-language-resource-endpoint>/language/analyze-text/
253249
"kind": "ExtractiveSummarization",
254250
"taskName": "Query_based Extractive Summarization",
255251
"parameters": {
256-
"query": "XYZ-code",
257-
"sentenceCount": 3
252+
"query": "XYZ-code"
258253
}
259254
}
260255
]
261256
}
262257
'
263258
```
264259

265-
### Using the summaryLength parameter
260+
### Summary length control
261+
262+
#### Using the summaryLength parameter in abstractive summarization
263+
264+
If you don't specify `summaryLength`, the model determines the summary length.
265+
266266
For the `summaryLength` parameter, three values are accepted:
267267
* oneSentence: Generates a summary of mostly 1 sentence, with around 80 tokens.
268268
* short: Generates a summary of mostly 2-3 sentences, with around 120 tokens.
269269
* medium: Generates a summary of mostly 4-6 sentences, with around 170 tokens.
270270
* long: Generates a summary of mostly over 7 sentences, with around 210 tokens.
271271

272+
Below is an example request:
273+
274+
```bash
275+
curl -i -X POST https://<your-language-resource-endpoint>/language/analyze-text/jobs?api-version=2023-04-01 \
276+
-H "Content-Type: application/json" \
277+
-H "Ocp-Apim-Subscription-Key: <your-language-resource-key>" \
278+
-d \
279+
'
280+
{
281+
"displayName": "Text Abstractive Summarization Task Example",
282+
"analysisInput": {
283+
"documents": [
284+
{
285+
"id": "1",
286+
"language": "en",
287+
"text": "At Microsoft, we have been on a quest to advance AI beyond existing techniques, by taking a more holistic, human-centric approach to learning and understanding. As Chief Technology Officer of Azure AI services, I have been working with a team of amazing scientists and engineers to turn this quest into a reality. In my role, I enjoy a unique perspective in viewing the relationship among three attributes of human cognition: monolingual text (X), audio or visual sensory signals, (Y) and multilingual (Z). At the intersection of all three, there’s magic—what we call XYZ-code as illustrated in Figure 1—a joint representation to create more powerful AI that can speak, hear, see, and understand humans better. We believe XYZ-code enables us to fulfill our long-term vision: cross-domain transfer learning, spanning modalities and languages. The goal is to have pretrained models that can jointly learn representations to support a broad range of downstream AI tasks, much in the way humans do today. Over the past five years, we have achieved human performance on benchmarks in conversational speech recognition, machine translation, conversational question answering, machine reading comprehension, and image captioning. These five breakthroughs provided us with strong signals toward our more ambitious aspiration to produce a leap in AI capabilities, achieving multi-sensory and multilingual learning that is closer in line with how humans learn and understand. I believe the joint XYZ-code is a foundational component of this aspiration, if grounded with external knowledge sources in the downstream AI tasks."
288+
}
289+
]
290+
},
291+
"tasks": [
292+
{
293+
"kind": "AbstractiveSummarization",
294+
"taskName": "Length controlled Abstractive Summarization",
295+
"parameters": {
296+
"sentenceLength": "short"
297+
}
298+
}
299+
]
300+
}
301+
'
302+
```
303+
304+
#### Using the sentenceCount parameter in extractive summarization
305+
For the `sentenceCount` parameter, you can input a value 1-20 to indicate the desired number of output sentences.
306+
307+
Below is an example request:
308+
309+
```bash
310+
curl -i -X POST https://<your-language-resource-endpoint>/language/analyze-text/jobs?api-version=2023-11-15-preview \
311+
-H "Content-Type: application/json" \
312+
-H "Ocp-Apim-Subscription-Key: <your-language-resource-key>" \
313+
-d \
314+
'
315+
{
316+
"displayName": "Text Extractive Summarization Task Example",
317+
"analysisInput": {
318+
"documents": [
319+
{
320+
"id": "1",
321+
"language": "en",
322+
"text": "At Microsoft, we have been on a quest to advance AI beyond existing techniques, by taking a more holistic, human-centric approach to learning and understanding. As Chief Technology Officer of Azure AI services, I have been working with a team of amazing scientists and engineers to turn this quest into a reality. In my role, I enjoy a unique perspective in viewing the relationship among three attributes of human cognition: monolingual text (X), audio or visual sensory signals, (Y) and multilingual (Z). At the intersection of all three, there’s magic—what we call XYZ-code as illustrated in Figure 1—a joint representation to create more powerful AI that can speak, hear, see, and understand humans better. We believe XYZ-code enables us to fulfill our long-term vision: cross-domain transfer learning, spanning modalities and languages. The goal is to have pretrained models that can jointly learn representations to support a broad range of downstream AI tasks, much in the way humans do today. Over the past five years, we have achieved human performance on benchmarks in conversational speech recognition, machine translation, conversational question answering, machine reading comprehension, and image captioning. These five breakthroughs provided us with strong signals toward our more ambitious aspiration to produce a leap in AI capabilities, achieving multi-sensory and multilingual learning that is closer in line with how humans learn and understand. I believe the joint XYZ-code is a foundational component of this aspiration, if grounded with external knowledge sources in the downstream AI tasks."
323+
}
324+
]
325+
},
326+
"tasks": [
327+
{
328+
"kind": "ExtractiveSummarization",
329+
"taskName": "Length controlled Extractive Summarization",
330+
"parameters": {
331+
"sentenceCount": "5"
332+
}
333+
}
334+
]
335+
}
336+
'
337+
```
338+
272339
## Service and data limits
273340

274341
[!INCLUDE [service limits article](../../includes/service-limits-link.md)]

articles/ai-services/language-service/summarization/region-support.md

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -25,11 +25,14 @@ Some summarization features are only available in limited regions. More regions
2525
|North Europe |&#9989; |&#9989; |&#10060; |
2626
|East US |&#9989; |&#9989; |&#9989; |
2727
|East US 2 |&#9989; |&#9989; |&#10060; |
28+
|Central US |&#9989; |&#9989; |&#10060; |
29+
|South Central US |&#9989; |&#9989; |&#10060; |
2830
|West US |&#9989; |&#9989; |&#10060; |
29-
|USNat West |&#9989; |&#9989; |&#10060; |
31+
|West US 2 |&#9989; |&#9989; |&#10060; |
3032
|USNat East |&#9989; |&#9989; |&#10060; |
31-
|USSec West |&#9989; |&#9989; |&#10060; |
33+
|USNat West |&#9989; |&#9989; |&#10060; |
3234
|USSec East |&#9989; |&#9989; |&#10060; |
35+
|USSec West |&#9989; |&#9989; |&#10060; |
3336
|South UK |&#9989; |&#9989; |&#10060; |
3437
|Southeast Asia |&#9989; |&#9989; |&#10060; |
3538
|Australia East |&#9989; |&#9989; |&#10060; |

articles/ai-services/openai/concepts/model-retirements.md

Lines changed: 9 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ titleSuffix: Azure OpenAI
44
description: Learn about the model deprecations and retirements in Azure OpenAI.
55
ms.service: azure-ai-openai
66
ms.topic: conceptual
7-
ms.date: 09/09/2024
7+
ms.date: 09/12/2024
88
ms.custom:
99
manager: nitinme
1010
author: mrbullwinkle
@@ -91,9 +91,9 @@ These models are currently available for use in Azure OpenAI Service.
9191

9292
| Model | Version | Retirement date | Suggested replacements |
9393
| ---- | ---- | ---- | --- |
94-
| `gpt-35-turbo` | 0301 | January 27, 2025<br><br> Deployments set to [**Auto-update to default**](/azure/ai-services/openai/how-to/working-with-models?tabs=powershell#auto-update-to-default) will be automatically upgraded to version: `0125`, starting on November 15, 2024. | `gpt-35-turbo` (0125) <br><br> `gpt-4o-mini` |
95-
| `gpt-35-turbo`<br>`gpt-35-turbo-16k` | 0613 | January 27, 2025 <br><br> Deployments set to [**Auto-update to default**](/azure/ai-services/openai/how-to/working-with-models?tabs=powershell#auto-update-to-default) will be automatically upgraded to version: `0125`, starting on November 15, 2024. | `gpt-35-turbo` (0125) <br><br> `gpt-4o-mini`|
96-
| `gpt-35-turbo` | 1106 | No earlier than Nov 17, 2024 <br><br> Deployments set to [**Auto-update to default**](/azure/ai-services/openai/how-to/working-with-models?tabs=powershell#auto-update-to-default) will be automatically upgraded to version: `0125`, starting on November 15, 2024. | `gpt-35-turbo` (0125) <br><br> `gpt-4o-mini` |
94+
| `gpt-35-turbo` | 0301 | January 27, 2025<br><br> Deployments set to [**Auto-update to default**](/azure/ai-services/openai/how-to/working-with-models?tabs=powershell#auto-update-to-default) will be automatically upgraded to version: `0125`, starting on November 13, 2024. | `gpt-35-turbo` (0125) <br><br> `gpt-4o-mini` |
95+
| `gpt-35-turbo`<br>`gpt-35-turbo-16k` | 0613 | January 27, 2025 <br><br> Deployments set to [**Auto-update to default**](/azure/ai-services/openai/how-to/working-with-models?tabs=powershell#auto-update-to-default) will be automatically upgraded to version: `0125`, starting on November 13, 2024. | `gpt-35-turbo` (0125) <br><br> `gpt-4o-mini`|
96+
| `gpt-35-turbo` | 1106 | No earlier than Nov 17, 2024 <br><br> Deployments set to [**Auto-update to default**](/azure/ai-services/openai/how-to/working-with-models?tabs=powershell#auto-update-to-default) will be automatically upgraded to version: `0125`, starting on November 13, 2024. | `gpt-35-turbo` (0125) <br><br> `gpt-4o-mini` |
9797
| `gpt-35-turbo` | 0125 | No earlier than Feb 22, 2025 | `gpt-4o-mini` |
9898
| `gpt-4`<br>`gpt-4-32k` | 0314 | June 6, 2025 | `gpt-4o` |
9999
| `gpt-4`<br>`gpt-4-32k` | 0613 | June 6, 2025 | `gpt-4o` |
@@ -115,7 +115,7 @@ These models are currently available for use in Azure OpenAI Service.
115115

116116
| Model | Current default version | New default version | Default upgrade date |
117117
|---|---|---|---|
118-
| `gpt-35-turbo` | 0301 | 0125 | Deployments of versions `0301`, `0613`, and `1106` set to [**Auto-update to default**](/azure/ai-services/openai/how-to/working-with-models?tabs=powershell#auto-update-to-default) will be automatically upgraded to version: `0125`, starting on November 15, 2024.|
118+
| `gpt-35-turbo` | 0301 | 0125 | Deployments of versions `0301`, `0613`, and `1106` set to [**Auto-update to default**](/azure/ai-services/openai/how-to/working-with-models?tabs=powershell#auto-update-to-default) will be automatically upgraded to version: `0125`, starting on November 13, 2024.|
119119

120120

121121

@@ -158,6 +158,10 @@ If you're an existing customer looking for information about these models, see [
158158

159159
## Retirement and deprecation history
160160

161+
## September 12, 2024
162+
163+
* `gpt-35-turbo` (0301), (0613), (1106) and `gpt-35-turbo-16k` (0613) auto-update to default upgrade date updated to November 13, 2024.
164+
161165
## September 9, 2024
162166

163167
* `gpt-35-turbo` (0301) and (0613) retirement changed to January 27, 2025.

0 commit comments

Comments
 (0)