Skip to content

Commit d6a900d

Browse files
authored
Merge pull request #240 from MicrosoftDocs/main
9/12 11:00 AM IST Publish
2 parents 022de22 + 3107ae2 commit d6a900d

File tree

8 files changed

+99
-27
lines changed

8 files changed

+99
-27
lines changed

articles/ai-services/language-service/summarization/how-to/document-summarization.md

Lines changed: 82 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -122,22 +122,11 @@ curl -i -X POST https://<your-language-resource-endpoint>/language/analyze-text/
122122
{
123123
"kind": "AbstractiveSummarization",
124124
"taskName": "Text Abstractive Summarization Task 1",
125-
"parameters": {
126-
"summaryLength": "short"
127-
}
128125
}
129126
]
130127
}
131128
'
132129
```
133-
If you don't specify `summaryLength`, the model determines the summary length.
134-
135-
### Using the summaryLength parameter
136-
For the `summaryLength` parameter, three values are accepted:
137-
* oneSentence: Generates a summary of mostly 1 sentence, with around 80 tokens.
138-
* short: Generates a summary of mostly 2-3 sentences, with around 120 tokens.
139-
* medium: Generates a summary of mostly 4-6 sentences, with around 170 tokens.
140-
* long: Generates a summary of mostly over 7 sentences, with around 210 tokens.
141130

142131
2. Make the following changes in the command where needed:
143132
- Replace the value `your-language-resource-key` with your key.
@@ -222,7 +211,14 @@ The following cURL commands are executed from a BASH shell. Edit these commands
222211

223212
The query-based text summarization API is an extension to the existing text summarization API.
224213

225-
The biggest difference is a new `query` field in the request body (under `tasks` > `parameters` > `query`). Additionally, there's a new way to specify the preferred `summaryLength` in "buckets" of short/medium/long, which we recommend using instead of `sentenceCount`, especially when using abstractive. Below is an example request:
214+
The biggest difference is a new `query` field in the request body (under `tasks` > `parameters` > `query`).
215+
216+
> [!TIP]
217+
> Query based summarization has some differentiation in the utilization of length control based on the type of query based summarization you're using:
218+
> - Query based extractive summarization supports length control by specifying sentenceCount.
219+
> - Query based abstractive summarization doesn't support length control.
220+
221+
Below is an example request:
226222

227223
```bash
228224
curl -i -X POST https://<your-language-resource-endpoint>/language/analyze-text/jobs?api-version=2023-11-15-preview \
@@ -253,22 +249,93 @@ curl -i -X POST https://<your-language-resource-endpoint>/language/analyze-text/
253249
"kind": "ExtractiveSummarization",
254250
"taskName": "Query_based Extractive Summarization",
255251
"parameters": {
256-
"query": "XYZ-code",
257-
"sentenceCount": 3
252+
"query": "XYZ-code"
258253
}
259254
}
260255
]
261256
}
262257
'
263258
```
264259

265-
### Using the summaryLength parameter
260+
### Summary length control
261+
262+
#### Using the summaryLength parameter in abstractive summarization
263+
264+
If you don't specify `summaryLength`, the model determines the summary length.
265+
266266
For the `summaryLength` parameter, three values are accepted:
267267
* oneSentence: Generates a summary of mostly 1 sentence, with around 80 tokens.
268268
* short: Generates a summary of mostly 2-3 sentences, with around 120 tokens.
269269
* medium: Generates a summary of mostly 4-6 sentences, with around 170 tokens.
270270
* long: Generates a summary of mostly over 7 sentences, with around 210 tokens.
271271

272+
Below is an example request:
273+
274+
```bash
275+
curl -i -X POST https://<your-language-resource-endpoint>/language/analyze-text/jobs?api-version=2023-04-01 \
276+
-H "Content-Type: application/json" \
277+
-H "Ocp-Apim-Subscription-Key: <your-language-resource-key>" \
278+
-d \
279+
'
280+
{
281+
"displayName": "Text Abstractive Summarization Task Example",
282+
"analysisInput": {
283+
"documents": [
284+
{
285+
"id": "1",
286+
"language": "en",
287+
"text": "At Microsoft, we have been on a quest to advance AI beyond existing techniques, by taking a more holistic, human-centric approach to learning and understanding. As Chief Technology Officer of Azure AI services, I have been working with a team of amazing scientists and engineers to turn this quest into a reality. In my role, I enjoy a unique perspective in viewing the relationship among three attributes of human cognition: monolingual text (X), audio or visual sensory signals, (Y) and multilingual (Z). At the intersection of all three, there’s magic—what we call XYZ-code as illustrated in Figure 1—a joint representation to create more powerful AI that can speak, hear, see, and understand humans better. We believe XYZ-code enables us to fulfill our long-term vision: cross-domain transfer learning, spanning modalities and languages. The goal is to have pretrained models that can jointly learn representations to support a broad range of downstream AI tasks, much in the way humans do today. Over the past five years, we have achieved human performance on benchmarks in conversational speech recognition, machine translation, conversational question answering, machine reading comprehension, and image captioning. These five breakthroughs provided us with strong signals toward our more ambitious aspiration to produce a leap in AI capabilities, achieving multi-sensory and multilingual learning that is closer in line with how humans learn and understand. I believe the joint XYZ-code is a foundational component of this aspiration, if grounded with external knowledge sources in the downstream AI tasks."
288+
}
289+
]
290+
},
291+
"tasks": [
292+
{
293+
"kind": "AbstractiveSummarization",
294+
"taskName": "Length controlled Abstractive Summarization",
295+
"parameters": {
296+
"sentenceLength": "short"
297+
}
298+
}
299+
]
300+
}
301+
'
302+
```
303+
304+
#### Using the sentenceCount parameter in extractive summarization
305+
For the `sentenceCount` parameter, you can input a value 1-20 to indicate the desired number of output sentences.
306+
307+
Below is an example request:
308+
309+
```bash
310+
curl -i -X POST https://<your-language-resource-endpoint>/language/analyze-text/jobs?api-version=2023-11-15-preview \
311+
-H "Content-Type: application/json" \
312+
-H "Ocp-Apim-Subscription-Key: <your-language-resource-key>" \
313+
-d \
314+
'
315+
{
316+
"displayName": "Text Extractive Summarization Task Example",
317+
"analysisInput": {
318+
"documents": [
319+
{
320+
"id": "1",
321+
"language": "en",
322+
"text": "At Microsoft, we have been on a quest to advance AI beyond existing techniques, by taking a more holistic, human-centric approach to learning and understanding. As Chief Technology Officer of Azure AI services, I have been working with a team of amazing scientists and engineers to turn this quest into a reality. In my role, I enjoy a unique perspective in viewing the relationship among three attributes of human cognition: monolingual text (X), audio or visual sensory signals, (Y) and multilingual (Z). At the intersection of all three, there’s magic—what we call XYZ-code as illustrated in Figure 1—a joint representation to create more powerful AI that can speak, hear, see, and understand humans better. We believe XYZ-code enables us to fulfill our long-term vision: cross-domain transfer learning, spanning modalities and languages. The goal is to have pretrained models that can jointly learn representations to support a broad range of downstream AI tasks, much in the way humans do today. Over the past five years, we have achieved human performance on benchmarks in conversational speech recognition, machine translation, conversational question answering, machine reading comprehension, and image captioning. These five breakthroughs provided us with strong signals toward our more ambitious aspiration to produce a leap in AI capabilities, achieving multi-sensory and multilingual learning that is closer in line with how humans learn and understand. I believe the joint XYZ-code is a foundational component of this aspiration, if grounded with external knowledge sources in the downstream AI tasks."
323+
}
324+
]
325+
},
326+
"tasks": [
327+
{
328+
"kind": "ExtractiveSummarization",
329+
"taskName": "Length controlled Extractive Summarization",
330+
"parameters": {
331+
"sentenceCount": "5"
332+
}
333+
}
334+
]
335+
}
336+
'
337+
```
338+
272339
## Service and data limits
273340

274341
[!INCLUDE [service limits article](../../includes/service-limits-link.md)]

articles/ai-services/language-service/summarization/region-support.md

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -25,11 +25,14 @@ Some summarization features are only available in limited regions. More regions
2525
|North Europe |&#9989; |&#9989; |&#10060; |
2626
|East US |&#9989; |&#9989; |&#9989; |
2727
|East US 2 |&#9989; |&#9989; |&#10060; |
28+
|Central US |&#9989; |&#9989; |&#10060; |
29+
|South Central US |&#9989; |&#9989; |&#10060; |
2830
|West US |&#9989; |&#9989; |&#10060; |
29-
|USNat West |&#9989; |&#9989; |&#10060; |
31+
|West US 2 |&#9989; |&#9989; |&#10060; |
3032
|USNat East |&#9989; |&#9989; |&#10060; |
31-
|USSec West |&#9989; |&#9989; |&#10060; |
33+
|USNat West |&#9989; |&#9989; |&#10060; |
3234
|USSec East |&#9989; |&#9989; |&#10060; |
35+
|USSec West |&#9989; |&#9989; |&#10060; |
3336
|South UK |&#9989; |&#9989; |&#10060; |
3437
|Southeast Asia |&#9989; |&#9989; |&#10060; |
3538
|Australia East |&#9989; |&#9989; |&#10060; |

articles/ai-services/speech-service/includes/how-to/professional-voice/create-consent/speech-studio.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,9 @@ To add a voice talent profile and upload their consent statement, follow these s
2727
1. In the **Add new voice talent** wizard, describe the characteristics of the voice you're going to create. The scenarios that you specify here must be consistent with what you provided in the application form.
2828
1. Select **Next**.
2929
1. On the **Upload voice talent statement** page, follow the instructions to upload the voice talent statement you've recorded beforehand. Make sure the verbal statement was [recorded](../../../../record-custom-voice-samples.md) with the same settings, environment, and speaking style as your training data.
30+
3031
:::image type="content" source="../../../../media/custom-voice/upload-verbal-statement.png" alt-text="Screenshot of the voice talent statement upload dialog.":::
32+
3133
1. Enter the voice talent name and company name. The voice talent name must be the name of the person who recorded the consent statement. Enter the name in the same language used in the recorded statement. The company name must match the company name that was spoken in the recorded statement. Ensure the company name is entered in the same language as the recorded statement.
3234
1. Select **Next**.
3335
1. Review the voice talent and persona details, and select **Submit**.

0 commit comments

Comments
 (0)