Skip to content

Commit 6cd81aa

Browse files
authored
Merge pull request #249846 from TimShererWithAquent/us141588g
Azure OpenAI Freshness Pass - User Story: 141588
2 parents 9240de4 + f2ac440 commit 6cd81aa

File tree

13 files changed

+736
-702
lines changed

13 files changed

+736
-702
lines changed

articles/ai-services/speech-service/how-to-speech-synthesis.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,15 +1,15 @@
11
---
22
title: "How to synthesize speech from text - Speech service"
33
titleSuffix: Azure AI services
4-
description: Learn how to convert text to speech. Learn about object construction and design patterns, supported audio output formats, and custom configuration options for speech synthesis.
4+
description: Learn how to convert text to speech, including object construction and design patterns, supported audio output formats, and custom configuration options.
55
services: cognitive-services
66
author: eric-urban
77
ms.author: eur
88
manager: nitinme
99
ms.service: cognitive-services
1010
ms.subservice: speech-service
1111
ms.topic: how-to
12-
ms.date: 09/16/2022
12+
ms.date: 08/30/2023
1313
ms.devlang: cpp, csharp, golang, java, javascript, objective-c, python
1414
ms.custom: devx-track-python, devx-track-js, devx-track-csharp, cog-serv-seo-aug-2020, mode-other, devx-track-extended-java, devx-track-go
1515
zone_pivot_groups: programming-languages-speech-services

articles/ai-services/speech-service/includes/how-to/speech-synthesis/cli.md

Lines changed: 14 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
author: eric-urban
33
ms.service: cognitive-services
44
ms.topic: include
5-
ms.date: 08/11/2020
5+
ms.date: 08/30/2023
66
ms.author: eur
77
---
88

@@ -18,27 +18,28 @@ ms.author: eur
1818

1919
## Synthesize speech to a speaker
2020

21-
Now you're ready to run the Speech CLI to synthesize speech from text. From the command line, change to the directory that contains the Speech CLI binary file. Then run the following command:
21+
Now you're ready to run the Speech CLI to synthesize speech from text.
2222

23-
```bash
24-
spx synthesize --text "I'm excited to try text to speech"
25-
```
23+
- In a console window, change to the directory that contains the Speech CLI binary file. Then run the following command:
2624

27-
The Speech CLI will produce natural language in English through the computer speaker.
25+
```console
26+
spx synthesize --text "I'm excited to try text to speech"
27+
```
28+
29+
The Speech CLI produces natural language in English through the computer speaker.
2830

2931
## Synthesize speech to a file
3032

31-
Run the following command to change the output from your speaker to a .wav file:
33+
- Run the following command to change the output from your speaker to a *.wav* file:
3234

33-
```bash
34-
spx synthesize --text "I'm excited to try text to speech" --audio output greetings.wav
35-
```
35+
```console
36+
spx synthesize --text "I'm excited to try text to speech" --audio output greetings.wav
37+
```
3638

37-
The Speech CLI will produce natural language in English in the `greetings.wav` audio file.
39+
The Speech CLI produces natural language in English to the *greetings.wav* audio file.
3840

3941
## Run and use a container
4042

4143
Speech containers provide websocket-based query endpoint APIs that are accessed through the Speech SDK and Speech CLI. By default, the Speech SDK and Speech CLI use the public Speech service. To use the container, you need to change the initialization method. Use a container host URL instead of key and region.
4244

43-
For more information about containers, see the [speech containers](../../../speech-container-howto.md#host-urls) how-to guide.
44-
45+
For more information about containers, see [Install and run Speech containers with Docker](../../../speech-container-howto.md).

articles/ai-services/speech-service/includes/how-to/speech-synthesis/cpp.md

Lines changed: 83 additions & 77 deletions
Large diffs are not rendered by default.

articles/ai-services/speech-service/includes/how-to/speech-synthesis/csharp.md

Lines changed: 71 additions & 63 deletions
Large diffs are not rendered by default.

articles/ai-services/speech-service/includes/how-to/speech-synthesis/events.md

Lines changed: 10 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -3,20 +3,20 @@ author: eric-urban
33
ms.service: cognitive-services
44
ms.subservice: speech-service
55
ms.topic: include
6-
ms.date: 11/14/2022
6+
ms.date: 08/30/2023
77
ms.author: eur
88
---
99

1010

1111
| Event | Description | Use case |
12-
| --- | --- | --- |
13-
|`BookmarkReached`|Signals that a bookmark was reached. To trigger a bookmark reached event, a `bookmark` element is required in the [SSML](../../../speech-synthesis-markup-structure.md#bookmark-element). This event reports the output audio's elapsed time between the beginning of synthesis and the `bookmark` element. The event's `Text` property is the string value that you set in the bookmark's `mark` attribute. The `bookmark` elements won't be spoken.|You can use the `bookmark` element to insert custom markers in SSML to get the offset of each marker in the audio stream. The `bookmark` element can be used to reference a specific location in the text or tag sequence.|
14-
|`SynthesisCanceled`|Signals that the speech synthesis was canceled.|You can confirm when synthesis has been canceled.|
15-
|`SynthesisCompleted`|Signals that speech synthesis has completed.|You can confirm when synthesis has completed.|
16-
|`SynthesisStarted`|Signals that speech synthesis has started.|You can confirm when synthesis has started.|
17-
|`Synthesizing`|Signals that speech synthesis is ongoing. This event fires each time the SDK receives an audio chunk from the Speech service.|You can confirm when synthesis is in progress.|
18-
|`VisemeReceived`|Signals that a viseme event was received.|[Visemes](../../../how-to-speech-synthesis-viseme.md) are often used to represent the key poses in observed speech. Key poses include the position of the lips, jaw, and tongue in producing a particular phoneme. You can use visemes to animate the face of a character as speech audio plays.|
19-
|`WordBoundary`|Signals that a word boundary was received. This event is raised at the beginning of each new spoken word, punctuation, and sentence. The event reports the current word's time offset (in ticks) from the beginning of the output audio. This event also reports the character position in the input text (or [SSML](../../../speech-synthesis-markup.md)) immediately before the word that's about to be spoken.|This event is commonly used to get relative positions of the text and corresponding audio. You might want to know about a new word, and then take action based on the timing. For example, you can get information that can help you decide when and for how long to highlight words as they're spoken.|
12+
|:--- |:--- |:--- |
13+
| `BookmarkReached` | Signals that a bookmark was reached. To trigger a bookmark reached event, a `bookmark` element is required in the [SSML](../../../speech-synthesis-markup-structure.md#bookmark-element). This event reports the output audio's elapsed time between the beginning of synthesis and the `bookmark` element. The event's `Text` property is the string value that you set in the bookmark's `mark` attribute. The `bookmark` elements aren't spoken. | You can use the `bookmark` element to insert custom markers in SSML to get the offset of each marker in the audio stream. The `bookmark` element can be used to reference a specific location in the text or tag sequence. |
14+
| `SynthesisCanceled` | Signals that the speech synthesis was canceled. | You can confirm when synthesis has been canceled. |
15+
| `SynthesisCompleted` | Signals that speech synthesis has completed. | You can confirm when synthesis has completed. |
16+
| `SynthesisStarted` | Signals that speech synthesis has started. | You can confirm when synthesis has started. |
17+
| `Synthesizing` | Signals that speech synthesis is ongoing. This event fires each time the SDK receives an audio chunk from the Speech service. | You can confirm when synthesis is in progress. |
18+
| `VisemeReceived` | Signals that a viseme event was received. | [Visemes](../../../how-to-speech-synthesis-viseme.md) are often used to represent the key poses in observed speech. Key poses include the position of the lips, jaw, and tongue in producing a particular phoneme. You can use visemes to animate the face of a character as speech audio plays. |
19+
| `WordBoundary` | Signals that a word boundary was received. This event is raised at the beginning of each new spoken word, punctuation, and sentence. The event reports the current word's time offset, in ticks, from the beginning of the output audio. This event also reports the character position in the input text or [SSML](../../../speech-synthesis-markup.md) immediately before the word that's about to be spoken. | This event is commonly used to get relative positions of the text and corresponding audio. You might want to know about a new word, and then take action based on the timing. For example, you can get information that can help you decide when and for how long to highlight words as they're spoken. |
2020

2121
> [!NOTE]
22-
> Events are raised as the output audio data becomes available, which will be faster than playback to an output device. The caller must appropriately synchronize streaming and real-time.
22+
> Events are raised as the output audio data becomes available, which is faster than playback to an output device. The caller must appropriately synchronize streaming and real-time.

0 commit comments

Comments
 (0)