You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-services/speech-service/includes/how-to/recognize-speech/cli.md
+4-5Lines changed: 4 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -2,7 +2,7 @@
2
2
author: eric-urban
3
3
ms.service: azure-ai-speech
4
4
ms.topic: include
5
-
ms.date: 09/01/2023
5
+
ms.date: 08/13/2024
6
6
ms.author: eur
7
7
---
8
8
@@ -21,11 +21,11 @@ spx recognize --microphone
21
21
> [!NOTE]
22
22
> The Speech CLI defaults to English. You can choose a different language [from the speech to text table](../../../../language-support.md?tabs=stt). For example, add `--source de-DE` to recognize German speech.
23
23
24
-
Speak into the microphone, and you can see transcription of your words into text in real-time. The Speech CLI stops after a period of silence, or when you select **Ctrl+C**.
24
+
Speak into the microphone, and you can see transcription of your words into text in realtime. The Speech CLI stops after a period of silence, or when you select **Ctrl+C**.
25
25
26
26
## Recognize speech from a file
27
27
28
-
The Speech CLI can recognize speech in many file formats and natural languages. In this example, you can use any *.wav* file (16 KHz or 8 KHz, 16-bit, and mono PCM) that contains English speech. Or if you want a quick sample, download the <ahref="https://github.com/Azure-Samples/cognitive-services-speech-sdk/blob/master/samples/csharp/sharedcontent/console/whatstheweatherlike.wav"download="whatstheweatherlike"target="_blank">whatstheweatherlike.wav <spanclass="docon docon-download x-hidden-focus"></span></a> file, and copy it to the same directory as the Speech CLI binary file.
28
+
The Speech CLI can recognize speech in many file formats and natural languages. In this example, you can use any *.wav* file (16 kHz or 8 kHz, 16-bit, and mono PCM) that contains English speech. Or if you want a quick sample, download the file [whatstheweatherlike.wav](https://github.com/Azure-Samples/cognitive-services-speech-sdk/blob/master/samples/csharp/sharedcontent/console/whatstheweatherlike.wav), and copy it to the same directory as the Speech CLI binary file.
29
29
30
30
Use the following command to run the Speech CLI to recognize speech found in the audio file:
31
31
@@ -42,5 +42,4 @@ The Speech CLI shows a text transcription of the speech on the screen.
42
42
43
43
Speech containers provide websocket-based query endpoint APIs that are accessed through the Speech SDK and Speech CLI. By default, the Speech SDK and Speech CLI use the public Speech service. To use the container, you need to change the initialization method. Use a container host URL instead of key and region.
44
44
45
-
For more information about containers, see [Host URLs](../../../speech-container-howto.md#host-urls) in Install and run Speech containers with Docker.
46
-
45
+
For more information about containers, see Host URLs in [Install and run Speech containers with Docker](../../../speech-container-howto.md#host-urls).
[`SetSpeechRecognitionLanguage`](/cpp/cognitive-services/speech/speechconfig#setspeechrecognitionlanguage) is a parameter that takes a string as an argument. For a list of supported locales, see [Language and voice support for the Speech service](../../../language-support.md?tabs=stt).
195
+
[`SetSpeechRecognitionLanguage`](/cpp/cognitive-services/speech/speechconfig#setspeechrecognitionlanguage) is a parameter that takes a string as an argument. For a list of supported locales, see [Language and voice support for the Speech service](../../../language-support.md).
196
196
197
197
## Language identification
198
198
199
199
You can use language identification with speech to text recognition when you need to identify the language in an audio source and then transcribe it to text.
200
200
201
-
For a complete code sample, see [Language identification](../../../language-identification.md?pivots=programming-language-cpp#use-speech-to-text).
201
+
For a complete code sample, see [Language identification](../../../language-identification.md?pivots=programming-language-cpp).
Copy file name to clipboardExpand all lines: articles/ai-services/speech-service/includes/how-to/recognize-speech/csharp.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -273,7 +273,7 @@ The [`SpeechRecognitionLanguage`](/dotnet/api/microsoft.cognitiveservices.speech
273
273
274
274
You can use language identification with speech to text recognition when you need to identify the language in an audio source and then transcribe it to text.
275
275
276
-
For a complete code sample, see [Language identification](../../../language-identification.md?pivots=programming-language-csharp#use-speech-to-text).
276
+
For a complete code sample, see [Language identification](../../../language-identification.md?pivots=programming-language-csharp).
Copy file name to clipboardExpand all lines: articles/ai-services/speech-service/includes/how-to/recognize-speech/java.md
+8-8Lines changed: 8 additions & 8 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -2,7 +2,7 @@
2
2
author: eric-urban
3
3
ms.service: azure-ai-speech
4
4
ms.topic: include
5
-
ms.date: 09/01/2023
5
+
ms.date: 08/13/2024
6
6
ms.custom: devx-track-java
7
7
ms.author: eur
8
8
---
@@ -15,8 +15,8 @@ ms.author: eur
15
15
16
16
To call the Speech service by using the Speech SDK, you need to create a [SpeechConfig](/java/api/com.microsoft.cognitiveservices.speech.speechconfig) instance. This class includes information about your subscription, like your key and associated region, endpoint, host, or authorization token.
17
17
18
-
1. Create a Speech resource in the [Azure portal](https://portal.azure.com/#create/Microsoft.CognitiveServicesSpeechServices).
19
-
1. Create a `SpeechConfig` instance by using your key and region.
18
+
1. Create a Speech resource in the [Azure portal](https://portal.azure.com/#create/Microsoft.CognitiveServicesSpeechServices). Get the Speech resource key and region.
19
+
1. Create a `SpeechConfig` instance by using your Speech key and region.
You can use language identification with speech to text recognition when you need to identify the language in an audio source and then transcribe it to text.
218
218
219
-
For a complete code sample, see [Language identification](../../../language-identification.md?pivots=programming-language-java#use-speech-to-text).
219
+
For a complete code sample, see [Language identification](../../../language-identification.md?pivots=programming-language-java).
220
220
221
221
## Use a custom endpoint
222
222
223
223
With [custom speech](../../../custom-speech-overview.md), you can upload your own data, test and train a custom model, compare accuracy between models, and deploy a model to a custom endpoint. The following example shows how to set a custom endpoint:
Copy file name to clipboardExpand all lines: articles/ai-services/speech-service/includes/how-to/recognize-speech/javascript.md
+15-14Lines changed: 15 additions & 14 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -2,7 +2,7 @@
2
2
author: eric-urban
3
3
ms.service: azure-ai-speech
4
4
ms.topic: include
5
-
ms.date: 09/01/2023
5
+
ms.date: 08/13/2024
6
6
ms.author: eur
7
7
ms.custom: devx-track-js
8
8
---
@@ -11,12 +11,12 @@ ms.custom: devx-track-js
11
11
12
12
[!INCLUDE [Introduction](intro.md)]
13
13
14
-
## Create a speech configuration
14
+
## Create a speech configuration instance
15
15
16
-
To call the Speech service by using the Speech SDK, you need to create a [`SpeechConfig`](/javascript/api/microsoft-cognitiveservices-speech-sdk/speechconfig) instance. This class includes information about your subscription, like your key and associated location/region, endpoint, host, or authorization token.
16
+
To call the Speech service by using the Speech SDK, you need to create a [`SpeechConfig`](/javascript/api/microsoft-cognitiveservices-speech-sdk/speechconfig) instance. This class includes information about your subscription, like your key and associated region, endpoint, host, or authorization token.
17
17
18
-
1. Create a `SpeechConfig` instance by using your key and location/region.
19
-
1. Create a Speech resource on the [Azure portal](https://portal.azure.com/#create/Microsoft.CognitiveServicesSpeechServices).
18
+
1. Create a Speech resource in the [Azure portal](https://portal.azure.com/#create/Microsoft.CognitiveServicesSpeechServices). Get the Speech resource key and region.
19
+
1. Create a `SpeechConfig` instance by using the following code. Replace `YourSpeechKey` and `YourSpeechRegion` with your Speech resource key and region.
@@ -36,7 +36,7 @@ You can initialize `SpeechConfig` in a few other ways:
36
36
Recognizing speech from a microphone isn't supported in Node.js. It's supported only in a browser-based JavaScript environment. For more information, see the [React sample](https://github.com/Azure-Samples/AzureSpeechReactSample) and the [implementation of speech to text from a microphone](https://github.com/Azure-Samples/AzureSpeechReactSample/blob/main/src/App.js#L29) on GitHub. The React sample shows design patterns for the exchange and management of authentication tokens. It also shows the capture of audio from a microphone or file for speech to text conversions.
37
37
38
38
> [!NOTE]
39
-
> If you want to use a *specific* audio input device, you need to specify the device ID in the `AudioConfig` object. For more information, see [Select an audio input device with the Speech SDK](../../../how-to-select-audio-input-devices.md).
39
+
> If you want to use a *specific* audio input device, you need to specify the device ID in `AudioConfig`. To learn how to get the device ID, see [Select an audio input device with the Speech SDK](../../../how-to-select-audio-input-devices.md).
40
40
41
41
## Recognize speech from a file
42
42
@@ -91,7 +91,7 @@ function fromStream() {
91
91
fromStream();
92
92
```
93
93
94
-
Using a push stream as input assumes that the audio data is a raw pulse-code modulation (PCM) data that skips any headers. The API still works in certain cases if the header wasn't skipped. For the best results, consider implementing logic to read off the headers so that `fs` begins at the *start of the audio data*.
94
+
Using a push stream as input assumes that the audio data is raw pulse-code modulation (PCM) data that skips any headers. The API still works in certain cases if the header isn't skipped. For the best results, consider implementing logic to read off the headers so that `fs` begins at the *start of the audio data*.
95
95
96
96
## Handle errors
97
97
@@ -126,7 +126,8 @@ switch (result.reason) {
126
126
127
127
The previous examples use single-shot recognition, which recognizes a single utterance. The end of a single utterance is determined by listening for silence at the end or until a maximum of 15 seconds of audio is processed.
128
128
129
-
In contrast, you can use continuous recognition when you want to control when to stop recognizing. It requires you to subscribe to the `Recognizing`, `Recognized`, and `Canceled` events to get the recognition results. To stop recognition, you must call [`stopContinuousRecognitionAsync`](/javascript/api/microsoft-cognitiveservices-speech-sdk/speechrecognizer#stopcontinuousrecognitionasync). Here's an example of how continuous recognition is performed on an audio input file.
129
+
In contrast, you can use continuous recognition when you want to control when to stop recognizing. It requires you to subscribe to the `Recognizing`, `Recognized`, and `Canceled` events to get the recognition results. To stop recognition, you must call [`stopContinuousRecognitionAsync`]
130
+
(/javascript/api/microsoft-cognitiveservices-speech-sdk/speechrecognizer#microsoft-cognitiveservices-speech-sdk-speechrecognizer-stopcontinuousrecognitionasync). Here's an example of how continuous recognition is performed on an audio input file.
130
131
131
132
Start by defining the input and initializing [`SpeechRecognizer`](/javascript/api/microsoft-cognitiveservices-speech-sdk/speechrecognizer):
With everything set up, call [`startContinuousRecognitionAsync`](/javascript/api/microsoft-cognitiveservices-speech-sdk/speechrecognizer#startcontinuousrecognitionasync) to start recognizing:
177
+
With everything set up, call [`startContinuousRecognitionAsync`]
178
+
(/javascript/api/microsoft-cognitiveservices-speech-sdk/speechrecognizer#microsoft-cognitiveservices-speech-sdk-speechrecognizer-startkeywordrecognitionasync) to start recognizing:
@@ -190,13 +192,13 @@ A common task for speech recognition is specifying the input (or source) languag
190
192
speechConfig.speechRecognitionLanguage="it-IT";
191
193
```
192
194
193
-
The [`speechRecognitionLanguage`](/javascript/api/microsoft-cognitiveservices-speech-sdk/speechconfig#speechrecognitionlanguage) property expects a language-locale format string. For more information, see the [list of supported speech to text locales](../../../language-support.md?tabs=stt).
195
+
The [`speechRecognitionLanguage`](/javascript/api/microsoft-cognitiveservices-speech-sdk/speechconfig#microsoft-cognitiveservices-speech-sdk-speechconfig-speechrecognitionlanguage) property expects a language-locale format string. For a list of supported locales, see [Language and voice support for the Speech service](../../../language-support.md).
194
196
195
197
## Language identification
196
198
197
-
You can use [language identification](../../../language-identification.md?pivots=programming-language-javascript#use-speech-to-text) with speech to text recognition when you need to identify the language in an audio source and then transcribe it to text.
199
+
You can use language identification with speech to text recognition when you need to identify the language in an audio source and then transcribe it to text.
198
200
199
-
For a complete code sample, see [Language identification](../../../language-identification.md?pivots=programming-language-javascript#use-speech-to-text).
201
+
For a complete code sample, see [Language identification](../../../language-identification.md?pivots=programming-language-javascript).
200
202
201
203
## Use a custom endpoint
202
204
@@ -212,5 +214,4 @@ var speechRecognizer = new SpeechSDK.SpeechRecognizer(speechConfig);
212
214
213
215
Speech containers provide websocket-based query endpoint APIs that are accessed through the Speech SDK and Speech CLI. By default, the Speech SDK and Speech CLI use the public Speech service. To use the container, you need to change the initialization method. Use a container host URL instead of key and region.
214
216
215
-
For more information about containers, see [Host URLs](../../../speech-container-howto.md#host-urls) in Install and run Speech containers with Docker.
216
-
217
+
For more information about containers, see Host URLs in [Install and run Speech containers with Docker](../../../speech-container-howto.md#host-urls).
Speech containers provide websocket-based query endpoint APIs that are accessed through the Speech SDK and Speech CLI. By default, the Speech SDK and Speech CLI use the public Speech service. To use the container, you need to change the initialization method. Use a container host URL instead of key and region.
36
36
37
-
For more information about containers, see [Host URLs](../../../speech-container-howto.md#host-urls) in Install and run Speech containers with Docker.
38
-
37
+
For more information about containers, see Host URLs in [Install and run Speech containers with Docker](../../../speech-container-howto.md#host-urls).
0 commit comments