Skip to content

Commit 6a29974

Browse files
authored
time offset for current word or sentence
1 parent 239a770 commit 6a29974

File tree

1 file changed

+2
-2
lines changed
  • articles/cognitive-services/Speech-Service/includes/how-to/speech-synthesis

1 file changed

+2
-2
lines changed

articles/cognitive-services/Speech-Service/includes/how-to/speech-synthesis/events.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ author: eric-urban
33
ms.service: cognitive-services
44
ms.subservice: speech-service
55
ms.topic: include
6-
ms.date: 01/25/2022
6+
ms.date: 11/14/2022
77
ms.author: eur
88
---
99

@@ -16,7 +16,7 @@ ms.author: eur
1616
|`SynthesisStarted`|Signals that speech synthesis has started.|You can confirm when synthesis has started.|
1717
|`Synthesizing`|Signals that speech synthesis is ongoing. This event fires each time the SDK receives an audio chunk from the Speech service.|You can confirm when synthesis is in progress.|
1818
|`VisemeReceived`|Signals that a viseme event was received.|[Visemes](../../../how-to-speech-synthesis-viseme.md) are often used to represent the key poses in observed speech. Key poses include the position of the lips, jaw, and tongue in producing a particular phoneme. You can use visemes to animate the face of a character as speech audio plays.|
19-
|`WordBoundary`|Signals that a word boundary was received. This event is raised at the beginning of each new spoken word, punctuation, and sentence. The event reports the output audio's elapsed time between the beginning of synthesis and the start of the next word. This event also reports the character position in the input text (or [SSML](../../../speech-synthesis-markup.md)) immediately before the word that's about to be spoken.|This event is commonly used to get relative positions of the text and corresponding audio. You might want to know about a new word, and then take action based on the timing. For example, you can get information that can help you decide when and for how long to highlight words as they're spoken.|
19+
|`WordBoundary`|Signals that a word boundary was received. This event is raised at the beginning of each new spoken word, punctuation, and sentence. The event reports the current word's time offset (in ticks) from the beginning of the output audio. This event also reports the character position in the input text (or [SSML](../../../speech-synthesis-markup.md)) immediately before the word that's about to be spoken.|This event is commonly used to get relative positions of the text and corresponding audio. You might want to know about a new word, and then take action based on the timing. For example, you can get information that can help you decide when and for how long to highlight words as they're spoken.|
2020

2121
> [!NOTE]
2222
> Events are raised as the output audio data becomes available, which will be faster than playback to an output device. The caller must appropriately synchronize streaming and real time.

0 commit comments

Comments
 (0)