Skip to content

Commit a1e308d

Browse files
committed
GA stt diarization
1 parent 2098f13 commit a1e308d

File tree

8 files changed

+9
-19
lines changed

8 files changed

+9
-19
lines changed

articles/ai-services/speech-service/get-started-stt-diarization.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ keywords: speech to text, speech to text software
1414
#customer intent: As a developer, I want to create speech to text applications that use diarization to improve readability of multiple person conversations.
1515
---
1616

17-
# Quickstart: Create real-time diarization (Preview)
17+
# Quickstart: Create real-time diarization
1818

1919
::: zone pivot="programming-language-csharp"
2020
[!INCLUDE [C# include](includes/quickstarts/stt-diarization/csharp.md)]

articles/ai-services/speech-service/includes/quickstarts/stt-diarization/cpp.md

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -142,9 +142,6 @@ Follow these steps to create a console application and install the Speech SDK.
142142

143143
The application recognizes speech from multiple participants in the conversation. Your audio file should contain multiple speakers.
144144

145-
> [!NOTE]
146-
> The service performs best with at least 7 seconds of continuous audio from a single speaker. This allows the system to differentiate the speakers properly. Otherwise the Speaker ID is returned as `Unknown`.
147-
148145
1. To change the speech recognition language, replace `en-US` with another [supported language](~/articles/cognitive-services/speech-service/supported-languages.md). For example, `es-ES` for Spanish (Spain). The default language is `en-US` if you don't specify a language. For details about how to identify one of multiple languages that might be spoken, see [language identification](~/articles/cognitive-services/speech-service/language-identification.md).
149146

150147
1. [Build and run](/cpp/build/vscpp-step-2-build) your application to start conversation transcription:

articles/ai-services/speech-service/includes/quickstarts/stt-diarization/csharp.md

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -121,9 +121,6 @@ Follow these steps to create a console application and install the Speech SDK.
121121
122122
The application recognizes speech from multiple participants in the conversation. Your audio file should contain multiple speakers.
123123
124-
> [!NOTE]
125-
> The service performs best with at least 7 seconds of continuous audio from a single speaker. This allows the system to differentiate the speakers properly. Otherwise the Speaker ID is returned as `Unknown`.
126-
127124
1. To change the speech recognition language, replace `en-US` with another [supported language](~/articles/cognitive-services/speech-service/supported-languages.md). For example, `es-ES` for Spanish (Spain). The default language is `en-US` if you don't specify a language. For details about how to identify one of multiple languages that might be spoken, see [language identification](~/articles/cognitive-services/speech-service/language-identification.md).
128125
129126
1. Run your console application to start conversation transcription:

articles/ai-services/speech-service/includes/quickstarts/stt-diarization/intro.md

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -8,9 +8,6 @@ ms.author: eur
88

99
In this quickstart, you run an application for speech to text transcription with real-time diarization. Diarization distinguishes between the different speakers who participate in the conversation. The Speech service provides information about which speaker was speaking a particular part of transcribed speech.
1010

11-
> [!NOTE]
12-
> Real-time diarization is currently in public preview.
13-
1411
The speaker information is included in the result in the speaker ID field. The speaker ID is a generic identifier assigned to each conversation participant by the service during the recognition as different speakers are being identified from the provided audio content.
1512

1613
> [!TIP]

articles/ai-services/speech-service/includes/quickstarts/stt-diarization/java.md

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -148,9 +148,6 @@ Follow these steps to create a console application for conversation transcriptio
148148

149149
The application recognizes speech from multiple participants in the conversation. Your audio file should contain multiple speakers.
150150

151-
> [!NOTE]
152-
> The service performs best with at least 7 seconds of continuous audio from a single speaker. This allows the system to differentiate the speakers properly. Otherwise the Speaker ID is returned as `Unknown`.
153-
154151
1. To change the speech recognition language, replace `en-US` with another [supported language](~/articles/cognitive-services/speech-service/supported-languages.md). For example, `es-ES` for Spanish (Spain). The default language is `en-US` if you don't specify a language. For details about how to identify one of multiple languages that might be spoken, see [language identification](~/articles/cognitive-services/speech-service/language-identification.md).
155152

156153
1. Run your new console application to start conversation transcription:

articles/ai-services/speech-service/includes/quickstarts/stt-diarization/javascript.md

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -93,9 +93,6 @@ Follow these steps to create a new console application for conversation transcri
9393

9494
The application recognizes speech from multiple participants in the conversation. Your audio file should contain multiple speakers.
9595

96-
> [!NOTE]
97-
> The service performs best with at least 7 seconds of continuous audio from a single speaker. This allows the system to differentiate the speakers properly. Otherwise the Speaker ID is returned as `Unknown`.
98-
9996
1. To change the speech recognition language, replace `en-US` with another [supported language](~/articles/cognitive-services/speech-service/supported-languages.md). For example, `es-ES` for Spanish (Spain). The default language is `en-US` if you don't specify a language. For details about how to identify one of multiple languages that might be spoken, see [language identification](~/articles/cognitive-services/speech-service/language-identification.md).
10097

10198
1. Run your new console application to start speech recognition from a file:

articles/ai-services/speech-service/includes/quickstarts/stt-diarization/python.md

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -108,9 +108,6 @@ Follow these steps to create a new console application.
108108

109109
The application recognizes speech from multiple participants in the conversation. Your audio file should contain multiple speakers.
110110

111-
> [!NOTE]
112-
> The service performs best with at least 7 seconds of continuous audio from a single speaker. This allows the system to differentiate the speakers properly. Otherwise the Speaker ID is returned as `Unknown`.
113-
114111
1. To change the speech recognition language, replace `en-US` with another [supported language](~/articles/cognitive-services/speech-service/supported-languages.md). For example, `es-ES` for Spanish (Spain). The default language is `en-US` if you don't specify a language. For details about how to identify one of multiple languages that might be spoken, see [language identification](~/articles/cognitive-services/speech-service/language-identification.md).
115112

116113
1. Run your new console application to start conversation transcription:

articles/ai-services/speech-service/includes/release-notes/release-notes-stt.md

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -6,6 +6,14 @@ ms.date: 3/13/2024
66
ms.author: eur
77
---
88

9+
### April 2024 release
10+
11+
#### Real-time speech to text with diariazation (GA)
12+
13+
Real-time speech to text with diariazation is now generally available.
14+
15+
Check out [Real-time diarization quickstart](../../get-started-stt-diarization.md) to learn more about how to create speech to text applications that use diarization to distinguish between the different speakers who participate in the conversation.
16+
917
### March 2024 release
1018

1119
#### Whisper general availability (GA)

0 commit comments

Comments
 (0)