Skip to content

Commit 39137a6

Browse files
committed
fix tabs and sample audio files
1 parent e2e49de commit 39137a6

File tree

5 files changed

+13
-13
lines changed

5 files changed

+13
-13
lines changed

articles/ai-services/speech-service/audio-processing-speech-sdk.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -123,7 +123,7 @@ MicrophoneCoordinates[] microphoneCoordinates = new MicrophoneCoordinates[7]
123123
};
124124
var microphoneArrayGeometry = new MicrophoneArrayGeometry(MicrophoneArrayType.Planar, microphoneCoordinates);
125125
var audioProcessingOptions = AudioProcessingOptions.Create(AudioProcessingConstants.AUDIO_INPUT_PROCESSING_ENABLE_DEFAULT, microphoneArrayGeometry, SpeakerReferenceChannel.LastChannel);
126-
var audioInput = AudioConfig.FromWavFileInput("katiesteve_mono.wav", audioProcessingOptions);
126+
var audioInput = AudioConfig.FromWavFileInput("katiesteve.wav", audioProcessingOptions);
127127

128128
var recognizer = new SpeechRecognizer(speechConfig, audioInput);
129129
```
@@ -139,7 +139,7 @@ MicrophoneArrayGeometry microphoneArrayGeometry
139139
{ { 0, 0, 0 }, { 40, 0, 0 }, { 20, -35, 0 }, { -20, -35, 0 }, { -40, 0, 0 }, { -20, 35, 0 }, { 20, 35, 0 } }
140140
};
141141
auto audioProcessingOptions = AudioProcessingOptions::Create(AUDIO_INPUT_PROCESSING_ENABLE_DEFAULT, microphoneArrayGeometry, SpeakerReferenceChannel::LastChannel);
142-
auto audioInput = AudioConfig::FromWavFileInput("katiesteve_mono.wav", audioProcessingOptions);
142+
auto audioInput = AudioConfig::FromWavFileInput("katiesteve.wav", audioProcessingOptions);
143143

144144
auto recognizer = SpeechRecognizer::FromConfig(speechConfig, audioInput);
145145
```
@@ -159,7 +159,7 @@ microphoneCoordinates[5] = new MicrophoneCoordinates(-20, 35, 0);
159159
microphoneCoordinates[6] = new MicrophoneCoordinates(20, 35, 0);
160160
MicrophoneArrayGeometry microphoneArrayGeometry = new MicrophoneArrayGeometry(MicrophoneArrayType.Planar, microphoneCoordinates);
161161
AudioProcessingOptions audioProcessingOptions = AudioProcessingOptions.create(AudioProcessingConstants.AUDIO_INPUT_PROCESSING_ENABLE_DEFAULT, microphoneArrayGeometry, SpeakerReferenceChannel.LastChannel);
162-
AudioConfig audioInput = AudioConfig.fromWavFileInput("katiesteve_mono.wav", audioProcessingOptions);
162+
AudioConfig audioInput = AudioConfig.fromWavFileInput("katiesteve.wav", audioProcessingOptions);
163163

164164
SpeechRecognizer recognizer = new SpeechRecognizer(speechConfig, audioInput);
165165
```

articles/ai-services/speech-service/includes/quickstarts/stt-diarization/cpp.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -59,7 +59,7 @@ Follow these steps to create a new console application and install the Speech SD
5959
6060
speechConfig->SetSpeechRecognitionLanguage("en-US");
6161
62-
auto audioConfig = AudioConfig::FromWavFileInput("katiesteve_mono.wav");
62+
auto audioConfig = AudioConfig::FromWavFileInput("katiesteve.wav");
6363
auto conversationTranscriber = ConversationTranscriber::FromConfig(speechConfig, audioConfig);
6464
6565
// promise for synchronization of recognition end.
@@ -134,7 +134,7 @@ Follow these steps to create a new console application and install the Speech SD
134134
}
135135
```
136136
137-
1. Replace `katiesteve_mono.wav` with the filepath and filename of your `.wav` file. The intent of this quickstart is to recognize speech from multiple participants in the conversation. Your audio file should contain multiple speakers. For example, you can use the [sample audio file](https://github.com/Azure-Samples/cognitive-services-speech-sdk/blob/master/quickstart/csharp/dotnet/conversation-transcription/helloworld/katiesteve_mono.wav) provided in the Speech SDK samples repository on GitHub.
137+
1. Replace `katiesteve.wav` with the filepath and filename of your `.wav` file. The intent of this quickstart is to recognize speech from multiple participants in the conversation. Your audio file should contain multiple speakers. For example, you can use the [sample audio file](https://github.com/Azure-Samples/cognitive-services-speech-sdk/blob/master/sampledata/audiofiles/katiesteve.wav) provided in the Speech SDK samples repository on GitHub.
138138
> [!NOTE]
139139
> The service performs best with at least 7 seconds of continuous audio from a single speaker. This allows the system to differentiate the speakers properly. Otherwise the Speaker ID is returned as `Unknown`.
140140
1. To change the speech recognition language, replace `en-US` with another [supported language](~/articles/cognitive-services/speech-service/supported-languages.md). For example, `es-ES` for Spanish (Spain). The default language is `en-US` if you don't specify a language. For details about how to identify one of multiple languages that might be spoken, see [language identification](~/articles/cognitive-services/speech-service/language-identification.md).

articles/ai-services/speech-service/includes/quickstarts/stt-diarization/csharp.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,7 @@ Follow these steps to create a new console application and install the Speech SD
4848
4949
async static Task Main(string[] args)
5050
{
51-
var filepath = "katiesteve_mono.wav";
51+
var filepath = "katiesteve.wav";
5252
var speechConfig = SpeechConfig.FromSubscription(speechKey, speechRegion);
5353
speechConfig.SpeechRecognitionLanguage = "en-US";
5454
@@ -110,7 +110,7 @@ Follow these steps to create a new console application and install the Speech SD
110110
}
111111
```
112112
113-
1. Replace `katiesteve_mono.wav` with the filepath and filename of your `.wav` file. The intent of this quickstart is to recognize speech from multiple participants in the conversation. Your audio file should contain multiple speakers. For example, you can use the [sample audio file](https://github.com/Azure-Samples/cognitive-services-speech-sdk/blob/master/quickstart/csharp/dotnet/conversation-transcription/helloworld/katiesteve_mono.wav) provided in the Speech SDK samples repository on GitHub.
113+
1. Replace `katiesteve.wav` with the filepath and filename of your `.wav` file. The intent of this quickstart is to recognize speech from multiple participants in the conversation. Your audio file should contain multiple speakers. For example, you can use the [sample audio file](https://github.com/Azure-Samples/cognitive-services-speech-sdk/blob/master/sampledata/audiofiles/katiesteve.wav) provided in the Speech SDK samples repository on GitHub.
114114
> [!NOTE]
115115
> The service performs best with at least 7 seconds of continuous audio from a single speaker. This allows the system to differentiate the speakers properly. Otherwise the Speaker ID is returned as `Unknown`.
116116
1. To change the speech recognition language, replace `en-US` with another [supported language](~/articles/cognitive-services/speech-service/supported-languages.md). For example, `es-ES` for Spanish (Spain). The default language is `en-US` if you don't specify a language. For details about how to identify one of multiple languages that might be spoken, see [language identification](~/articles/cognitive-services/speech-service/language-identification.md).

articles/ai-services/speech-service/includes/quickstarts/stt-diarization/java.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -82,7 +82,7 @@ Follow these steps to create a new console application for conversation transcri
8282

8383
SpeechConfig speechConfig = SpeechConfig.fromSubscription(speechKey, speechRegion);
8484
speechConfig.setSpeechRecognitionLanguage("en-US");
85-
AudioConfig audioInput = AudioConfig.fromWavFileInput("katiesteve_mono.wav");
85+
AudioConfig audioInput = AudioConfig.fromWavFileInput("katiesteve.wav");
8686

8787
Semaphore stopRecognitionSemaphore = new Semaphore(0);
8888

@@ -139,7 +139,7 @@ Follow these steps to create a new console application for conversation transcri
139139
}
140140
```
141141

142-
1. Replace `katiesteve_mono.wav` with the filepath and filename of your `.wav` file. The intent of this quickstart is to recognize speech from multiple participants in the conversation. Your audio file should contain multiple speakers. For example, you can use the [sample audio file](https://github.com/Azure-Samples/cognitive-services-speech-sdk/blob/master/quickstart/csharp/dotnet/conversation-transcription/helloworld/katiesteve_mono.wav) provided in the Speech SDK samples repository on GitHub.
142+
1. Replace `katiesteve.wav` with the filepath and filename of your `.wav` file. The intent of this quickstart is to recognize speech from multiple participants in the conversation. Your audio file should contain multiple speakers. For example, you can use the [sample audio file](https://github.com/Azure-Samples/cognitive-services-speech-sdk/blob/master/sampledata/audiofiles/katiesteve.wav) provided in the Speech SDK samples repository on GitHub.
143143
> [!NOTE]
144144
> The service performs best with at least 7 seconds of continuous audio from a single speaker. This allows the system to differentiate the speakers properly. Otherwise the Speaker ID is returned as `Unknown`.
145145
1. To change the speech recognition language, replace `en-US` with another [supported language](~/articles/cognitive-services/speech-service/supported-languages.md). For example, `es-ES` for Spanish (Spain). The default language is `en-US` if you don't specify a language. For details about how to identify one of multiple languages that might be spoken, see [language identification](~/articles/cognitive-services/speech-service/language-identification.md).

articles/ai-services/speech-service/includes/quickstarts/stt-diarization/python.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -64,7 +64,7 @@ Follow these steps to create a new console application.
6464
speech_config = speechsdk.SpeechConfig(subscription=os.environ.get('SPEECH_KEY'), region=os.environ.get('SPEECH_REGION'))
6565
speech_config.speech_recognition_language="en-US"
6666

67-
audio_config = speechsdk.audio.AudioConfig(filename="katiesteve_mono.wav")
67+
audio_config = speechsdk.audio.AudioConfig(filename="katiesteve.wav")
6868
conversation_transcriber = speechsdk.transcription.ConversationTranscriber(speech_config=speech_config, audio_config=audio_config)
6969

7070
transcribing_stop = False
@@ -100,7 +100,7 @@ Follow these steps to create a new console application.
100100
print("Encountered exception. {}".format(err))
101101
```
102102

103-
1. Replace `katiesteve_mono.wav` with the filepath and filename of your `.wav` file. The intent of this quickstart is to recognize speech from multiple participants in the conversation. Your audio file should contain multiple speakers. For example, you can use the [sample audio file](https://github.com/Azure-Samples/cognitive-services-speech-sdk/blob/master/quickstart/csharp/dotnet/conversation-transcription/helloworld/katiesteve_mono.wav) provided in the Speech SDK samples repository on GitHub.
103+
1. Replace `katiesteve.wav` with the filepath and filename of your `.wav` file. The intent of this quickstart is to recognize speech from multiple participants in the conversation. Your audio file should contain multiple speakers. For example, you can use the [sample audio file](https://github.com/Azure-Samples/cognitive-services-speech-sdk/blob/master/sampledata/audiofiles/katiesteve.wav) provided in the Speech SDK samples repository on GitHub.
104104
> [!NOTE]
105105
> The service performs best with at least 7 seconds of continuous audio from a single speaker. This allows the system to differentiate the speakers properly. Otherwise the Speaker ID is returned as `Unknown`.
106106
1. To change the speech recognition language, replace `en-US` with another [supported language](~/articles/cognitive-services/speech-service/supported-languages.md). For example, `es-ES` for Spanish (Spain). The default language is `en-US` if you don't specify a language. For details about how to identify one of multiple languages that might be spoken, see [language identification](~/articles/cognitive-services/speech-service/language-identification.md).
@@ -140,9 +140,9 @@ TRANSCRIBED:
140140
Text=That's exciting. Let me try it right now.
141141
Speaker ID=Guest-2
142142
Canceled event
143-
CLOSING on ConversationTranscriptionCanceledEventArgs(session_id=606e8b5e65b94419b824d224127d9f92, result=ConversationTranscriptionResult(result_id=21d17c5738b442f8a7d428d0d5363fa8, speaker_id=, text=, reason=ResultReason.Canceled))
143+
CLOSING on ConversationTranscriptionCanceledEventArgs(session_id=92a0abb68636471dac07041b335d9be3, result=ConversationTranscriptionResult(result_id=ad1b1d83b5c742fcacca0692baa8df74, speaker_id=, text=, reason=ResultReason.Canceled))
144144
SessionStopped event
145-
CLOSING on SessionEventArgs(session_id=606e8b5e65b94419b824d224127d9f92)
145+
CLOSING on SessionEventArgs(session_id=92a0abb68636471dac07041b335d9be3)
146146
```
147147

148148
Speakers are identified as Guest-1, Guest-2, and so on, depending on the number of speakers in the conversation.

0 commit comments

Comments
 (0)