You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/cognitive-services/Speech-Service/faq-tts.yml
+6-6Lines changed: 6 additions & 6 deletions
Original file line number
Diff line number
Diff line change
@@ -26,7 +26,7 @@ sections:
26
26
- question: |
27
27
What audio formats does Text-to-Speech support?
28
28
answer: |
29
-
The Speech service supports 48-kHz, 24-kHz, 16-kHz, and 8-kHz audio outputs. Each prebuilt neural voice model is available at 24kHz and high-fidelity 48kHz. See [Audio outputs](rest-text-to-speech?tabs=streaming#audio-outputs).
29
+
The Speech service supports 48-kHz, 24-kHz, 16-kHz, and 8-kHz audio outputs. Each prebuilt neural voice model is available at 24kHz and high-fidelity 48kHz. See [Audio outputs](rest-text-to-speech.md?tabs=streaming#audio-outputs).
30
30
- question: |
31
31
How can we balance dynamic and static content to limit cost?
32
32
answer: |
@@ -38,7 +38,7 @@ sections:
38
38
- question: |
39
39
Is there a mapping between Viseme IDs and mouth shape?
40
40
answer: |
41
-
Yes. See [Get facial position with viseme](how-to-speech-synthesis-viseme?tabs=visemeid#map-phonemes-to-visemes).
41
+
Yes. See [Get facial position with viseme](how-to-speech-synthesis-viseme.md?tabs=visemeid#map-phonemes-to-visemes).
42
42
- question: |
43
43
Can the Visemes be mapped to UE5 MetaHuman blend shapes weights?
44
44
answer: |
@@ -53,7 +53,7 @@ sections:
53
53
- question: |
54
54
Can we have multiple strength for each emotions, like very sad, slightly sad and so on in?
55
55
answer: |
56
-
Adjusting the style degree is supported for some voices depending on the locale. See the [mstts:express-as tag](speech-synthesis-markup-voice#speaking-styles-and-roles).
56
+
Adjusting the style degree is supported for some voices depending on the locale. See the [mstts:express-as tag](speech-synthesis-markup-voice.md#speaking-styles-and-roles).
57
57
58
58
- name: Custom Neural Voice
59
59
questions:
@@ -84,7 +84,7 @@ sections:
84
84
- question: |
85
85
Is the model version the same as the engine version?
86
86
answer: |
87
-
No. The model version is different from the engine version. The model version means the version of the training recipe for your model and varies by the features supported and model training time. Azure Cognitive Services text-to-speech engines are updated from time to time to capture the latest language model that defines the pronunciation of the language. After you've trained your voice, you can apply your voice to the new language model by updating to the latest engine version. When a new engine is available, you're prompted to update your neural voice model. See [Update engine version for your voice model](how-to-custom-voice-create-voice?tabs=neural#update-engine-version-for-your-voice-model).
87
+
No. The model version is different from the engine version. The model version means the version of the training recipe for your model and varies by the features supported and model training time. Azure Cognitive Services text-to-speech engines are updated from time to time to capture the latest language model that defines the pronunciation of the language. After you've trained your voice, you can apply your voice to the new language model by updating to the latest engine version. When a new engine is available, you're prompted to update your neural voice model. See [Update engine version for your voice model](how-to-custom-voice-create-voice.md?tabs=neural#update-engine-version-for-your-voice-model).
88
88
- question: |
89
89
What kind of script should be prepared for a domain specific scenario such as gaming?
90
90
answer: |
@@ -96,7 +96,7 @@ sections:
96
96
- question: |
97
97
Switching styles via SSML only works for prebuilt neural voices, right?
98
98
answer: |
99
-
Switching styles via SSML is only for prebuilt multi-style voices. Custom Neural Voice does support [multi-style training](how-to-custom-voice-create-voice?tabs=multistyle#train-your-custom-neural-voice-model) for the same model, so you can also adjust the styles via SSML. Only for the speaking styles you have created for CNV.
99
+
Switching styles via SSML is only for prebuilt multi-style voices. Custom Neural Voice does support [multi-style training](how-to-custom-voice-create-voice.md?tabs=multistyle#train-your-custom-neural-voice-model) for the same model, so you can also adjust the styles via SSML. Only for the speaking styles you have created for CNV.
100
100
- question: |
101
101
Is it correct that after one training we can't train again unless we upload a corpus file?
102
102
answer: |
@@ -128,7 +128,7 @@ sections:
128
128
- question: |
129
129
Do we need to return the written permission from the voice talent back to Microsoft?
130
130
answer: |
131
-
Microsoft doesn't need the written permission, but you must obtain consent from your voice talent. The voice talent will also be required to record the consent statement and it must be uploaded into Speech Studio before training can begin. See [Set up voice talent for Custom Neural Voice](how-to-custom-voice-talent).
131
+
Microsoft doesn't need the written permission, but you must obtain consent from your voice talent. The voice talent will also be required to record the consent statement and it must be uploaded into Speech Studio before training can begin. See [Set up voice talent for Custom Neural Voice](how-to-custom-voice-talent.md).
0 commit comments