Skip to content

Commit 6ef3400

Browse files
committed
Merge branch 'main' of https://github.com/MicrosoftDocs/azure-docs-pr into us297578-content-curation
2 parents d31ca7c + c787984 commit 6ef3400

File tree

18 files changed

+974
-850
lines changed

18 files changed

+974
-850
lines changed

.openpublishing.redirection.json

Lines changed: 10 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -4544,6 +4544,16 @@
45444544
"source_path_from_root": "/articles/virtual-network/template-samples.md",
45454545
"redirect_url": "/samples/browse/?expanded=azure&products=azure-resource-manager&terms=virtual%20network",
45464546
"redirect_document_id": false
4547+
},
4548+
{
4549+
"source_path_from_root": "/articles/virtual-network/tutorial-restrict-network-access-to-resources-cli.md",
4550+
"redirect_url": "/azure/virtual-network/tutorial-restrict-network-access-to-resources",
4551+
"redirect_document_id": false
4552+
},
4553+
{
4554+
"source_path_from_root": "/articles/virtual-network/tutorial-restrict-network-access-to-resources-powershell.md",
4555+
"redirect_url": "/azure/virtual-network/tutorial-restrict-network-access-to-resources",
4556+
"redirect_document_id": false
45474557
}
45484558
]
45494559
}

articles/ai-services/speech-service/how-to-recognize-speech.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ author: eric-urban
66
manager: nitinme
77
ms.service: azure-ai-speech
88
ms.topic: how-to
9-
ms.date: 1/21/2024
9+
ms.date: 08/13/2024
1010
ms.author: eur
1111
ms.devlang: cpp
1212
ms.custom: devx-track-extended-java, devx-track-go, devx-track-js, devx-track-python
@@ -56,7 +56,7 @@ keywords: speech to text, speech to text software
5656
[!INCLUDE [CLI include](includes/how-to/recognize-speech/cli.md)]
5757
::: zone-end
5858

59-
## Next steps
59+
## Related content
6060

6161
* [Try the speech to text quickstart](get-started-speech-to-text.md)
6262
* [Improve recognition accuracy with custom speech](custom-speech-overview.md)

articles/ai-services/speech-service/includes/how-to/recognize-speech/cli.md

Lines changed: 4 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
author: eric-urban
33
ms.service: azure-ai-speech
44
ms.topic: include
5-
ms.date: 09/01/2023
5+
ms.date: 08/13/2024
66
ms.author: eur
77
---
88

@@ -21,11 +21,11 @@ spx recognize --microphone
2121
> [!NOTE]
2222
> The Speech CLI defaults to English. You can choose a different language [from the speech to text table](../../../../language-support.md?tabs=stt). For example, add `--source de-DE` to recognize German speech.
2323
24-
Speak into the microphone, and you can see transcription of your words into text in real-time. The Speech CLI stops after a period of silence, or when you select **Ctrl+C**.
24+
Speak into the microphone, and you can see transcription of your words into text in real time. The Speech CLI stops after a period of silence, or when you select **Ctrl+C**.
2525

2626
## Recognize speech from a file
2727

28-
The Speech CLI can recognize speech in many file formats and natural languages. In this example, you can use any *.wav* file (16 KHz or 8 KHz, 16-bit, and mono PCM) that contains English speech. Or if you want a quick sample, download the <a href="https://github.com/Azure-Samples/cognitive-services-speech-sdk/blob/master/samples/csharp/sharedcontent/console/whatstheweatherlike.wav" download="whatstheweatherlike" target="_blank">whatstheweatherlike.wav <span class="docon docon-download x-hidden-focus"></span></a> file, and copy it to the same directory as the Speech CLI binary file.
28+
The Speech CLI can recognize speech in many file formats and natural languages. In this example, you can use any *.wav* file (16 kHz or 8 kHz, 16-bit, and mono PCM) that contains English speech. Or if you want a quick sample, download the file [whatstheweatherlike.wav](https://github.com/Azure-Samples/cognitive-services-speech-sdk/blob/master/samples/csharp/sharedcontent/console/whatstheweatherlike.wav), and copy it to the same directory as the Speech CLI binary file.
2929

3030
Use the following command to run the Speech CLI to recognize speech found in the audio file:
3131

@@ -42,5 +42,4 @@ The Speech CLI shows a text transcription of the speech on the screen.
4242

4343
Speech containers provide websocket-based query endpoint APIs that are accessed through the Speech SDK and Speech CLI. By default, the Speech SDK and Speech CLI use the public Speech service. To use the container, you need to change the initialization method. Use a container host URL instead of key and region.
4444

45-
For more information about containers, see [Host URLs](../../../speech-container-howto.md#host-urls) in Install and run Speech containers with Docker.
46-
45+
For more information about containers, see Host URLs in [Install and run Speech containers with Docker](../../../speech-container-howto.md#host-urls).

articles/ai-services/speech-service/includes/how-to/recognize-speech/cpp.md

Lines changed: 12 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -2,20 +2,20 @@
22
author: eric-urban
33
ms.service: azure-ai-speech
44
ms.topic: include
5-
ms.date: 09/01/2023
5+
ms.date: 08/13/2024
66
ms.author: eur
77
---
88

99
[!INCLUDE [Header](../../common/cpp.md)]
1010

1111
[!INCLUDE [Introduction](intro.md)]
1212

13-
## Create a speech configuration
13+
## Create a speech configuration instance
1414

15-
To call the Speech service using the Speech SDK, you need to create a [`SpeechConfig`](/cpp/cognitive-services/speech/speechconfig) instance. This class includes information about your subscription, like your key and associated location/region, endpoint, host, or authorization token.
15+
To call the Speech service using the Speech SDK, you need to create a [`SpeechConfig`](/cpp/cognitive-services/speech/speechconfig) instance. This class includes information about your subscription, like your key and associated region, endpoint, host, or authorization token.
1616

17-
1. Create a `SpeechConfig` instance by using your key and region.
18-
1. Create a Speech resource on the [Azure portal](https://portal.azure.com/#create/Microsoft.CognitiveServicesSpeechServices).
17+
1. Create a Speech resource in the [Azure portal](https://portal.azure.com/#create/Microsoft.CognitiveServicesSpeechServices). Get the Speech resource key and region.
18+
1. Create a `SpeechConfig` instance by using the following code. Replace `YourSpeechKey` and `YourSpeechRegion` with your Speech resource key and region.
1919

2020
```cpp
2121
using namespace std;
@@ -48,11 +48,11 @@ auto result = speechRecognizer->RecognizeOnceAsync().get();
4848
cout << "RECOGNIZED: Text=" << result->Text << std::endl;
4949
```
5050

51-
If you want to use a *specific* audio input device, you need to specify the device ID in `AudioConfig`. For more information on how to get the device ID for your audio input device, see [Select an audio input device with the Speech SDK](../../../how-to-select-audio-input-devices.md)
51+
If you want to use a *specific* audio input device, you need to specify the device ID in `AudioConfig`. To learn how to get the device ID, see [Select an audio input device with the Speech SDK](../../../how-to-select-audio-input-devices.md).
5252

5353
## Recognize speech from a file
5454

55-
If you want to recognize speech from an audio file instead of using a microphone, you still need to create an `AudioConfig` instance. But for this case you don't call `FromDefaultMicrophoneInput()`. You call `FromWavFileInput()` and pass the file path:
55+
If you want to recognize speech from an audio file instead of using a microphone, you still need to create an `AudioConfig` instance. However, you don't call `FromDefaultMicrophoneInput()`. You call `FromWavFileInput()` and pass the file path:
5656

5757
```cpp
5858
using namespace Microsoft::CognitiveServices::Speech::Audio;
@@ -110,7 +110,7 @@ switch (result->Reason)
110110

111111
## Continuous recognition
112112

113-
Continuous recognition is a bit more involved than single-shot recognition. It requires you to subscribe to the `Recognizing`, `Recognized`, and `Canceled` events to get the recognition results. To stop recognition, you must call [StopContinuousRecognitionAsync](/cpp/cognitive-services/speech/speechrecognizer#stopcontinuousrecognitionasync). Here's an example of how continuous recognition is performed on an audio input file.
113+
Continuous recognition is a bit more involved than single-shot recognition. It requires you to subscribe to the `Recognizing`, `Recognized`, and `Canceled` events to get the recognition results. To stop recognition, you must call [StopContinuousRecognitionAsync](/cpp/cognitive-services/speech/speechrecognizer#stopcontinuousrecognitionasync). Here's an example of continuous recognition performed on an audio input file.
114114

115115
Start by defining the input and initializing [`SpeechRecognizer`](/cpp/cognitive-services/speech/speechrecognizer):
116116

@@ -192,13 +192,13 @@ A common task for speech recognition is specifying the input (or source) languag
192192
speechConfig->SetSpeechRecognitionLanguage("de-DE");
193193
```
194194
195-
[`SetSpeechRecognitionLanguage`](/cpp/cognitive-services/speech/speechconfig#setspeechrecognitionlanguage) is a parameter that takes a string as an argument. For more information, see the [list of supported speech to text locales](../../../language-support.md?tabs=stt).
195+
[`SetSpeechRecognitionLanguage`](/cpp/cognitive-services/speech/speechconfig#setspeechrecognitionlanguage) is a parameter that takes a string as an argument. For a list of supported locales, see [Language and voice support for the Speech service](../../../language-support.md).
196196
197197
## Language identification
198198
199-
You can use [language identification](../../../language-identification.md?pivots=programming-language-cpp#use-speech-to-text) with speech to text recognition when you need to identify the language in an audio source and then transcribe it to text.
199+
You can use language identification with speech to text recognition when you need to identify the language in an audio source and then transcribe it to text.
200200
201-
For a complete code sample, see [Language identification](../../../language-identification.md?pivots=programming-language-cpp#use-speech-to-text).
201+
For a complete code sample, see [Language identification](../../../language-identification.md?pivots=programming-language-cpp).
202202
203203
## Use a custom endpoint
204204
@@ -214,5 +214,4 @@ auto speechRecognizer = SpeechRecognizer::FromConfig(speechConfig);
214214

215215
Speech containers provide websocket-based query endpoint APIs that are accessed through the Speech SDK and Speech CLI. By default, the Speech SDK and Speech CLI use the public Speech service. To use the container, you need to change the initialization method. Use a container host URL instead of key and region.
216216

217-
For more information about containers, see [Host URLs](../../../speech-container-howto.md#host-urls) in Install and run Speech containers with Docker.
218-
217+
For more information about containers, see Host URLs in [Install and run Speech containers with Docker](../../../speech-container-howto.md#host-urls).

0 commit comments

Comments
 (0)