Skip to content

Commit 7005e18

Browse files
Merge pull request #223643 from eric-urban/eur/maintenance-pass
content maintenance
2 parents bc69ca8 + 59af5d7 commit 7005e18

File tree

4 files changed

+7
-10
lines changed

4 files changed

+7
-10
lines changed

articles/cognitive-services/Speech-Service/batch-synthesis.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ This diagram provides a high-level overview of the workflow.
2626
![Diagram of the Batch Synthesis API workflow.](media/long-audio-api/long-audio-api-workflow.png)
2727

2828
> [!TIP]
29-
> You can also use the [Speech SDK](speech-sdk.md) to create synthesized audio longer than 10 minutes by iterating over the text and synthesizing it in chunks.
29+
> You can also use the [Speech SDK](speech-sdk.md) to create synthesized audio longer than 10 minutes by iterating over the text and synthesizing it in chunks. For a C# example, see [GitHub](https://github.com/Azure-Samples/cognitive-services-speech-sdk/blob/master/samples/csharp/sharedcontent/console/speech_synthesis_samples.cs).
3030
3131
You can use the following REST API operations for batch synthesis:
3232

articles/cognitive-services/Speech-Service/ingestion-client.md

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -41,12 +41,11 @@ Internally, the tool uses Speech and Language services, and follows best practic
4141

4242
:::image type="content" source="media/ingestion-client/architecture-1.png" alt-text="Diagram that shows the Ingestion Client Architecture.":::
4343

44-
The following Speech service features are used by the Ingestion Client:
44+
The following Speech service feature is used by the Ingestion Client:
4545

4646
- [Batch speech-to-text](./batch-transcription.md): Transcribe large amounts of audio files asynchronously including speaker diarization and is typically used in post-call analytics scenarios. Diarization is the process of recognizing and separating speakers in mono channel audio data.
47-
- [Speaker identification](./speaker-recognition-overview.md): Helps you determine an unknown speaker’s identity within a group of enrolled speakers and is typically used for call center customer verification scenarios or fraud detection.
4847

49-
Language service features used by the Ingestion Client:
48+
Here are some Language service features that are used by the Ingestion Client:
5049

5150
- [Personally Identifiable Information (PII) extraction and redaction](../language-service/personally-identifiable-information/how-to-call-for-conversations.md): Identify, categorize, and redact sensitive information in conversation transcription.
5251
- [Sentiment analysis and opinion mining](../language-service/sentiment-opinion-mining/overview.md): Analyze transcriptions and associate positive, neutral, or negative sentiment at the utterance and conversation-level.

articles/cognitive-services/Speech-Service/overview.md

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -42,12 +42,13 @@ Speech feature summaries are provided below with links for more information.
4242

4343
Use [speech-to-text](speech-to-text.md) to transcribe audio into text, either in real time or asynchronously.
4444

45+
> [!TIP]
46+
> You can try speech-to-text in [Speech Studio](https://aka.ms/speechstudio/speechtotexttool) without signing up or writing any code.
47+
4548
Convert audio to text from a range of sources, including microphones, audio files, and blob storage. Use speaker diarisation to determine who said what and when. Get readable transcripts with automatic formatting and punctuation.
4649

4750
The base model may not be sufficient if the audio contains ambient noise or includes a lot of industry and domain-specific jargon. In these cases, you can create and train [custom speech models](custom-speech-overview.md) with acoustic, language, and pronunciation data. Custom speech models are private and can offer a competitive advantage.
4851

49-
You can try speech to text with [this demo web app](https://azure.microsoft.com/services/cognitive-services/speech-to-text/#features) or in the [Speech Studio](https://aka.ms/speechstudio/speechtotexttool).
50-
5152
### Text-to-speech
5253

5354
With [text to speech](text-to-speech.md), you can convert input text into humanlike synthesized speech. Use neural voices, which are humanlike voices powered by deep neural networks. Use the [Speech Synthesis Markup Language (SSML)](speech-synthesis-markup.md) to fine-tune the pitch, pronunciation, speaking rate, volume, and more.
@@ -101,8 +102,6 @@ We offer quickstarts in many popular programming languages. Each quickstart is d
101102
* [Speech-to-text quickstart](get-started-speech-to-text.md)
102103
* [Text-to-speech quickstart](get-started-text-to-speech.md)
103104
* [Speech translation quickstart](./get-started-speech-translation.md)
104-
* [Intent recognition quickstart](./get-started-intent-recognition.md)
105-
* [Speaker recognition quickstart](./get-started-speaker-recognition.md)
106105

107106
## Code samples
108107

articles/cognitive-services/Speech-Service/speech-to-text.md

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -21,15 +21,14 @@ In this overview, you learn about the benefits and capabilities of the speech-to
2121
Speech-to-text, also known as speech recognition, enables real-time or offline transcription of audio streams into text. For a full list of available speech-to-text languages, see [Language and voice support for the Speech service](language-support.md?tabs=stt).
2222

2323
> [!NOTE]
24-
> Microsoft uses the same recognition technology for Cortana and Office products.
24+
> Microsoft uses the same recognition technology for Windows and Office products.
2525
2626
## Get started
2727

2828
To get started, try the [speech-to-text quickstart](get-started-speech-to-text.md). Speech-to-text is available via the [Speech SDK](speech-sdk.md), the [REST API](rest-speech-to-text.md), and the [Speech CLI](spx-overview.md).
2929

3030
In depth samples are available in the [Azure-Samples/cognitive-services-speech-sdk](https://aka.ms/csspeech/samples) repository on GitHub. There are samples for C# (including UWP, Unity, and Xamarin), C++, Java, JavaScript (including Browser and Node.js), Objective-C, Python, and Swift. Code samples for Go are available in the [Microsoft/cognitive-services-speech-sdk-go](https://github.com/Microsoft/cognitive-services-speech-sdk-go) repository on GitHub.
3131

32-
3332
## Batch transcription
3433

3534
Batch transcription is a set of [Speech-to-text REST API](rest-speech-to-text.md) operations that enable you to transcribe a large amount of audio in storage. You can point to audio files with a shared access signature (SAS) URI and asynchronously receive transcription results. For more information on how to use the batch transcription API, see [How to use batch transcription](batch-transcription.md) and [Batch transcription samples (REST)](https://github.com/Azure-Samples/cognitive-services-speech-sdk/tree/master/samples/batch).

0 commit comments

Comments
 (0)