You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-services/speech-service/includes/language-support/multilingual-voices.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -16,4 +16,4 @@ ms.author: v-baolianzou
16
16
17
17
<sup>2</sup> The neural voice is a multilingual voice in Azure AI Speech. All multilingual voices can speak in the language in default locale of the input text without [using SSML](../../speech-synthesis-markup-voice.md#adjust-speaking-languages). However, you can still use the `<lang xml:lang>` element to adjust the speaking accent of each language to set preferred accent such as British accent (`en-GB`) for English. Check the [full list](https://speech.microsoft.com/portal/voicegallery) of supported locales through SSML.
18
18
19
-
<sup>3</sup> The OpenAI text to speech voices in Azure AI Speech are in public preview and only available in North Central US (`northcentralus`) and Sweden Central (`swedencentral`). Locales not listed for OpenAI voices aren't supported by design.
19
+
<sup>3</sup> The OpenAI text to speech voices in Azure AI Speech are in public preview and only available in North Central US (`northcentralus`) and Sweden Central (`swedencentral`). Locales not listed for OpenAI voices aren't supported. For information about additional differences between OpenAI text to speech voices and Azure AI Speech text to speech voices, see [OpenAI text to speech voices](../../openai-voices.md#openai-text-to-speech-voices-via-azure-openai-service-or-via-azure-ai-speech).
Copy file name to clipboardExpand all lines: articles/ai-services/speech-service/includes/language-support/tts.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -165,5 +165,5 @@ ms.custom: references_regions
165
165
166
166
<sup>3</sup> The neural voice is a multilingual voice in Azure AI Speech.
167
167
168
-
<sup>4</sup> The OpenAI text to speech voices in Azure AI Speech are in public preview and only available in North Central US (`northcentralus`) and Sweden Central (`swedencentral`).
168
+
<sup>4</sup> The OpenAI text to speech voices in Azure AI Speech are in public preview and only available in North Central US (`northcentralus`) and Sweden Central (`swedencentral`). Locales not listed for OpenAI voices aren't supported. For information about additional differences between OpenAI text to speech voices and Azure AI Speech text to speech voices, see [OpenAI text to speech voices](../../openai-voices.md#openai-text-to-speech-voices-via-azure-openai-service-or-via-azure-ai-speech).
Copy file name to clipboardExpand all lines: articles/ai-services/speech-service/openai-voices.md
+8-3Lines changed: 8 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -7,7 +7,7 @@ ms.author: eur
7
7
manager: nitinme
8
8
ms.service: azure-ai-speech
9
9
ms.topic: overview
10
-
ms.date: 2/1/2024
10
+
ms.date: 4/23/2024
11
11
ms.reviewer: v-baolianzou
12
12
ms.custom: references_regions
13
13
#customer intent: As a user who implements text to speech, I want to understand the options and differences between available OpenAI text to speech voices in Azure AI services.
@@ -55,13 +55,18 @@ Here's a comparison of features between OpenAI text to speech voices in Azure Op
55
55
|**Real-time or batch synthesis**| Real-time | Real-time and batch synthesis | Real-time and batch synthesis |
56
56
|**Latency**| greater than 500 ms | greater than 500 ms | less than 300 ms |
57
57
|**Sample rate of synthesized audio**| 24 kHz | 8, 16, 24, and 48 kHz | 8, 16, 24, and 48 kHz |
There are additional features and capabilities available in Azure AI Speech that aren't available with OpenAI voices. For example:
61
+
- OpenAI text to speech voices in Azure AI Speech [only support a subset of SSML elements](#ssml-elements-supported-by-openai-text-to-speech-voices-in-azure-ai-speech). Azure AI Speech voices support the full set of SSML elements.
62
+
- Azure AI Speech supports [word boundary events](./how-to-speech-synthesis.md#subscribe-to-synthesizer-events). OpenAI voices don't support word boundary events.
63
+
59
64
60
65
## SSML elements supported by OpenAI text to speech voices in Azure AI Speech
61
66
62
67
The [Speech Synthesis Markup Language (SSML)](./speech-synthesis-markup.md) with input text determines the structure, content, and other characteristics of the text to speech output. For example, you can use SSML to define a paragraph, a sentence, a break or a pause, or silence. You can wrap text with event tags such as bookmark or viseme that can be processed later by your application.
63
68
64
-
The following table outlines the Speech Synthesis Markup Language (SSML) elements supported by OpenAI text to speech voices in Azure AI speech. Only a subset of SSML tags are supported for OpenAI voices. See [SSML document structure and events](speech-synthesis-markup-structure.md) for more information.
69
+
The following table outlines the Speech Synthesis Markup Language (SSML) elements supported by OpenAI text to speech voices in Azure AI speech. Only the following subset of SSML tags are supported for OpenAI voices. See [SSML document structure and events](speech-synthesis-markup-structure.md) for more information.
0 commit comments