You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
To install the C++ Speech SDK from the binaries, download the SDK as a <ahref="https://aka.ms/csspeech/linuxbinary"target="_blank">.tar package <spanclass="docon docon-navigate-external x-hidden-focus"></span></a> and unpack the files in a directory of your choice. The following table shows the SDK folder structure (which include header files for both x86 and x64 target architectures):
31
+
Alternatively, the C++ Speech SDK can be installed from binaries. Download the SDK as a <ahref="https://aka.ms/csspeech/linuxbinary"target="_blank">.tar package <spanclass="docon docon-navigate-external x-hidden-focus"></span></a> and unpack the files in a directory of your choice. The contents of this package (which include header files for both x86 and x64 target architectures) are structured as follows:
Copy file name to clipboardExpand all lines: articles/cognitive-services/Speech-Service/includes/get-speech-sdk-dotnet.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -8,7 +8,7 @@ ms.author: dapine
8
8
9
9
:::row:::
10
10
:::column span="3":::
11
-
The .NET Speech SDK is available as a NuGet package and implements .NET Standard 2.0, for more information see <ahref="https://www.nuget.org/packages/Microsoft.CognitiveServices.Speech"target="_blank">Microsoft.CognitiveServices.Speech <spanclass="docon docon-navigate-external x-hidden-focus"></span></a>.
11
+
The .NET Speech SDK is available as a NuGet package and implements .NET Standard 2.0, for more information, see <ahref="https://www.nuget.org/packages/Microsoft.CognitiveServices.Speech"target="_blank">Microsoft.CognitiveServices.Speech <spanclass="docon docon-navigate-external x-hidden-focus"></span></a>.
12
12
:::column-end:::
13
13
:::column:::
14
14
<br>
@@ -20,13 +20,13 @@ ms.author: dapine
20
20
21
21
#### C# NuGet Package
22
22
23
-
To install the .NET Speech SDK from the **.NET Core CLI**run the following command.
23
+
The .NET Speech SDK can be installed from the **.NET Core CLI**with the following`dotnet add` command.
Copy file name to clipboardExpand all lines: articles/cognitive-services/Speech-Service/includes/get-speech-sdk-ios.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -8,7 +8,7 @@ ms.author: dapine
8
8
9
9
:::row:::
10
10
:::column span="3":::
11
-
When developing for iOS, there are two Speech SDKs available. The Objective-C Speech SDK is available natively as an iOS CocoaPod package, or the .NET Speech SDK could be used with Xamarin.iOS as it implements .NET Standard 2.0.
11
+
When developing for iOS, there are two Speech SDKs available. The Objective-C Speech SDK is available natively as an iOS CocoaPod package. Alternatively, the .NET Speech SDK could be used with Xamarin.iOS as it implements .NET Standard 2.0.
Copy file name to clipboardExpand all lines: articles/cognitive-services/Speech-Service/includes/get-speech-sdk-linux.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -23,7 +23,7 @@ ms.author: dapine
23
23
24
24
### System requirements
25
25
26
-
For a native applications, the Speech SDK relies on `libMicrosoft.CognitiveServices.Speech.core.so`. Make sure the target architecture (x86, x64) matches the application. Depending on the Linux version, additional dependencies may be required.
26
+
For a native application, the Speech SDK relies on `libMicrosoft.CognitiveServices.Speech.core.so`. Make sure the target architecture (x86, x64) matches the application. Depending on the Linux version, additional dependencies may be required.
27
27
28
28
- The shared libraries of the GNU C library (including the POSIX Threads Programming library, `libpthreads`)
29
29
- The OpenSSL library (`libssl.so.1.0.0` or `libssl.so.1.0.2`)
Copy file name to clipboardExpand all lines: articles/cognitive-services/Speech-Service/includes/get-speech-sdk-python.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -8,7 +8,7 @@ ms.author: dapine
8
8
9
9
:::row:::
10
10
:::column span="3":::
11
-
The Python Speech SDK is available as a Python Package Index (PyPI) module, for more information see <ahref="https://pypi.org/project/azure-cognitiveservices-speech/"target="_blank">azure-cognitiveservices-speech <spanclass="docon docon-navigate-external x-hidden-focus"></span></a>. The Python Speech SDK is compatible with Windows, Linux, and macOS.
11
+
The Python Speech SDK is available as a Python Package Index (PyPI) module, for more information, see <ahref="https://pypi.org/project/azure-cognitiveservices-speech/"target="_blank">azure-cognitiveservices-speech <spanclass="docon docon-navigate-external x-hidden-focus"></span></a>. The Python Speech SDK is compatible with Windows, Linux, and macOS.
Copy file name to clipboardExpand all lines: articles/cognitive-services/Speech-Service/includes/get-speech-sdk-windows.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -8,7 +8,7 @@ ms.author: dapine
8
8
9
9
:::row:::
10
10
:::column span="3":::
11
-
The Speech SDK supports Windows 10 and Windows Server 2016, or later versions. Earlier versions are **not** officially supported, however; it is possible to use parts of the Speech SDK with earlier versions of Windows.
11
+
The Speech SDK supports Windows 10 and Windows Server 2016, or later versions. Earlier versions are **not** officially supported. It is possible to use parts of the Speech SDK with earlier versions of Windows, although it's not advised.
Copy file name to clipboardExpand all lines: articles/cognitive-services/Speech-Service/speech-sdk.md
+7-7Lines changed: 7 additions & 7 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -14,29 +14,29 @@ ms.author: dapine
14
14
15
15
# About the Speech SDK
16
16
17
-
The Speech software development kit (SDK) exposes many of the Speech service capabilities, making it easier to develop speech-enabled applications. There are various SDKs available in many programming languages. All of the Speech SDKs are cross-platform, with the exception of the Objective-C / Swift SDK (which is only available on iOS and macOS).
17
+
The Speech software development kit (SDK) exposes many of the Speech service capabilities, making it easier to develop speech-enabled applications. There are various SDKs available in many programming languages. All of the Speech SDKs are cross-platform, with the exception of the Objective-C SDK (which is only available on iOS and macOS).
The Speech SDK exposes many features from the Speech service, but not all of them. The capabilities of the Speech SDK are often associated to scenarios. It's ideal for both real-time and non-real-time scenarios, utilizing local devices, files, and even input and output streams. There are [known limitations](#known-limitations) with the Speech SDK, where feature gaps exist. When a scenario is unachievable with the Speech SDK, look for a REST API alternative.
23
+
The Speech SDK exposes many features from the Speech service, but not all of them. The capabilities of the Speech SDK are often associated to scenarios. It's ideal for both real-time and non-real-time scenarios, using local devices, files, and even input and output streams. There are [known limitations](#known-limitations) with the Speech SDK, where feature gaps exist. When a scenario is unachievable with the Speech SDK, look for a REST API alternative.
24
24
25
25
### Speech-to-text
26
26
27
27
Speech-to-text (also known as *speech recognition*) transcribes audio streams to text that your applications, tools, or devices can consume or display. Use speech-to-text with [Language Understanding (LUIS)](https://docs.microsoft.com/azure/cognitive-services/luis) to derive user intents from transcribed speech and act on voice commands. For more information, see [Speech-to-text basics](speech-to-text-basics.md).
28
28
29
29
### Text-to-speech
30
30
31
-
Text-to-speech (also known as *speech synthesis*) converts input text into human-like synthesized speech using the [Speech Synthesis Markup Language (SSML)](speech-synthesis-markup.md). Choose from standard or neural voices, for more information, see [Text-to-speech language and voice support](language-support.md#text-to-speech).
31
+
Text-to-speech (also known as *speech synthesis*) converts text into human-like synthesized speech, using the [Speech Synthesis Markup Language (SSML)](speech-synthesis-markup.md). For more information on standard or neural voices, see [Text-to-speech language and voice support](language-support.md#text-to-speech).
32
32
33
33
### Keyword spotting
34
34
35
35
The concept of [keyword spotting](speech-devices-sdk-create-kws.md) is supported in the Speech SDK. Keyword spotting is the act of identifying a keyword in speech, followed by an action upon hearing the keyword. For example, "Hey Cortana" would activate the Cortana assistant.
36
36
37
37
### Voice assistants
38
38
39
-
Voice assistants using the Speech service empower developers to create natural, human-like conversational interfaces for their applications and experiences. The voice assistant service provides fast, reliable interaction between a device and an assistant implementation that uses the Bot Framework's Direct Line Speech channel or the integrated Custom Commands (Preview) service for task completion.
39
+
Voice assistants using the Speech SDK enable developers to create natural, human-like conversational interfaces for their applications and experiences. The voice assistant service provides fast, reliable interaction between a device and an assistant. The implementation uses the Bot Framework's Direct Line Speech channel or the integrated Custom Commands (Preview) service for task completion.
40
40
41
41
### Meeting scenarios
42
42
@@ -64,15 +64,15 @@ Several of the Speech SDKs' support codec compressed audio input streams. For mo
64
64
65
65
## Known limitations
66
66
67
-
While the Speech SDK covers many feature capabilities with various scenarios, there are known limitations. Certain functionalities are only available from the Azure portalor the REST API. An example of this is endpoint management. There are several other limitations to consider.
67
+
While the Speech SDK covers many feature capabilities with various scenarios, there are known limitations. Certain functionalities are only available from the Azure portal, Custom Speech portal, Custom voice portal, or the REST API. As an example, endpoint management is not possible through the Speech SDK.
68
68
69
69
### Batch transcription
70
70
71
-
Batch transcription enables asynchronous speech-to-text transcription of large volumes of data. This is a REST-based service however, which uses the same endpoint as customization and model management. Batch transcription is only possible from the REST API.
71
+
Batch transcription enables asynchronous speech-to-text transcription of large volumes of data. It is a REST-based service however, which uses the same endpoint as customization and model management. Batch transcription is only possible from the REST API.
72
72
73
73
### Custom Speech-to-text
74
74
75
-
If you are using speech-to-text for recognition and transcription in a unique environment, you can create and train custom acoustic, language, and pronunciation models to address ambient noise or industry-specific vocabulary. The creation and management of Custom Speech models is only available through the [Custom Speech Portal](https://aka.ms/customspeech), and not the Speech SDK. However, once the Custom Speech model is published it can be consumed by the Speech SDK.
75
+
When using speech-to-text for recognition and transcription in a unique environment, you can create and train custom acoustic, language, and pronunciation models to address ambient noise or industry-specific vocabulary. The creation and management of Custom Speech models is only available through the [Custom Speech Portal](https://aka.ms/customspeech), and not the Speech SDK. However, once the Custom Speech model is published it can be consumed by the Speech SDK.
0 commit comments