You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/cognitive-services/Speech-Service/speech-sdk.md
+11-79Lines changed: 11 additions & 79 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -14,7 +14,7 @@ ms.author: dapine
14
14
15
15
# About the Speech SDK
16
16
17
-
The Speech software development kit (SDK) exposes many of the Speech service capabilities, to empower you to develop speech-enabled applications. The Speech SDK is available in many programming languages, all of which work cross-platform, except for Objective-C, which is only available on iOS and macOS.
17
+
The Speech software development kit (SDK) exposes many of the Speech service capabilities, to empower you to develop speech-enabled applications. The Speech SDK is available in many programming languages and across all platforms.
@@ -56,21 +56,24 @@ The Speech SDK can be used for transcribing call center scenarios, where telepho
56
56
57
57
#### Call Center Transcription
58
58
59
-
A common scenario for speech-to-text is transcribing large volumes of telephony data that may come from various systems, such as Interactive Voice Response (IVR). The latest speech recognition models from the Speech service excel at transcribing this telephony data, even in cases when the data is difficult for a human to understand.
59
+
[Call Center Transcription](call-center-transcription.md) is common scenario for speech-to-text for transcribing large volumes of telephony data that may come from various systems, such as Interactive Voice Response (IVR). The latest speech recognition models from the Speech service excel at transcribing this telephony data, even in cases when the data is difficult for a human to understand.
60
60
61
61
### Codec compressed audio input
62
62
63
63
Several of the Speech SDK programming languages support codec compressed audio input streams. For more information, see <ahref="https://docs.microsoft.com/azure/cognitive-services/speech-service/how-to-use-codec-compressed-audio-input-streams"target="_blank">use compressed audio input formats <spanclass="docon docon-navigate-external x-hidden-focus"></span></a>.
64
64
65
65
## REST API
66
66
67
-
While the Speech SDK covers many feature capabilities of the Speech Service, for some scenarios you might want to use the REST API. Certain functionalities are only available from the Azure portal, Custom Speech portal, Custom Voice portal, or the REST API. As an example, endpoint management is only exposed via the REST API.
67
+
While the Speech SDK covers many feature capabilities of the Speech Service, for some scenarios you might want to use the REST API. As an example, endpoint management is only exposed via the REST API.
68
68
69
69
> [!TIP]
70
-
> When relying on the REST API, use the <ahref="https://editor.swagger.io/"target="_blank">Swagger Editor <spanclass="docon docon-navigate-external x-hidden-focus"></span></a> to automatically generate client libraries.
71
-
> For example, to generate a Batch transcription client library:
> When relying on the REST API, use the Swagger Editor to automatically generate client libraries. For example, to generate a Batch transcription client library.
71
+
> 1. Copy the example URL below:
72
+
> ```http
73
+
> https://westus.cris.ai/docs/v2.0/swagger
74
+
> ```
75
+
> 1. Navigate to the <a href="https://editor.swagger.io/" target="_blank">Swagger Editor <span class="docon docon-navigate-external x-hidden-focus"></span></a>
76
+
> 1. Select **File** > **Import URL** and paste the URL
74
77
> 1. Select **Generate Client** and choose your desired programming language
75
78
76
79
### Batch transcription
@@ -89,78 +92,7 @@ When using speech-to-text for recognition and transcription in a unique environm
89
92
90
93
Custom text-to-speech, also known as Custom Voice is a set of online tools that allow you to create a recognizable, one-of-a-kind voice for your brand. The creation and management of no-code Custom Voice models is available through the [Custom Voice Portal](https://aka.ms/customvoice). Once the Custom Voice model is published, it can be consumed by the Speech SDK.
91
94
92
-
## Get the SDK
93
-
94
-
# [Windows](#tab/windows)
95
-
96
-
> [!WARNING]
97
-
> The Speech SDK supports Windows 10 and Windows Server 2016, or later versions. Earlier versions are **not supported**.
98
-
99
-
The Speech SDK requires the <ahref="https://support.microsoft.com/help/2977003/the-latest-supported-visual-c-downloads"target="_blank">Microsoft Visual C++ Redistributable for Visual Studio 2019 <spanclass="docon docon-navigate-external x-hidden-focus"></span></a> on the system.
100
-
101
-
- <ahref="https://aka.ms/vs/16/release/vc_redist.x86.exe"target="_blank">Install for x86 <spanclass="docon docon-navigate-external x-hidden-focus"></span></a>
102
-
- <ahref="https://aka.ms/vs/16/release/vc_redist.x64.exe"target="_blank">Install for x64 <spanclass="docon docon-navigate-external x-hidden-focus"></span></a>
103
-
- <ahref="https://aka.ms/vs/16/release/vc_redist.arm64.exe"target="_blank">Install for ARMx64 <spanclass="docon docon-navigate-external x-hidden-focus"></span></a>
104
-
105
-
For microphone input, the Media Foundation libraries must be installed. These libraries are part of Windows 10 and Windows Server 2016. It's possible to use the Speech SDK without these libraries, as long as a microphone isn't used as the audio input device.
106
-
107
-
The required Speech SDK files can be deployed in the same directory as your application. This way your application can directly access the libraries. Make sure you select the correct version (x86/x64) that matches your application.
|`Microsoft.CognitiveServices.Speech.core.dll`| Core SDK, required for native and managed deployment |
112
-
|`Microsoft.CognitiveServices.Speech.csharp.dll`| Required for managed deployment |
113
-
114
-
> [!NOTE]
115
-
> Starting with the release 1.3.0 the file `Microsoft.CognitiveServices.Speech.csharp.bindings.dll` (shipped in previous releases) isn't needed anymore. The functionality is now integrated in the core SDK.
116
-
117
-
> [!NOTE]
118
-
> For the Windows Forms App (.NET Framework) C# project, make sure the libraries are included in your project's deployment settings. You can check this under `Properties -> Publish Section`. Click the `Application Files` button and find corresponding libraries from the scroll down list. Make sure the value is set to `Included`. Visual Studio will include the file when project is published/deployed.
119
-
120
-
For Windows, we support the following languages:
121
-
122
-
* C# (UWP and .NET), C++:
123
-
You can reference and use the latest version of our Speech SDK NuGet package. The package includes 32-bit and 64-bit client libraries and managed (.NET) libraries. The SDK can be installed in Visual Studio by using NuGet, [Microsoft.CognitiveServices.Speech](https://www.nuget.org/packages/Microsoft.CognitiveServices.Speech).
124
-
125
-
* Java:
126
-
You can reference and use the latest version of our Speech SDK Maven package, which supports only Windows x64. In your Maven project, add `https://csspeechstorage.blob.core.windows.net/maven/` as an additional repository and reference `com.microsoft.cognitiveservices.speech:client-sdk:1.8.0` as a dependency.
127
-
128
-
# [Linux](#tab/linux)
129
-
130
-
> [!NOTE]
131
-
> Currently, we only support the following distributions and development languages/platforms:
> **[<aname="footnote1">1</a>]** Linux ARM64 requires .NET Core 3.x (dotnet-sdk-3.x package) for proper ARM64 support.<br>
153
-
> **[<aname="footnote2">2</a>]** Follow the instructions on [how to configure RHEL/CentOS 7 for Speech SDK](~/articles/cognitive-services/speech-service/how-to-configure-rhel-centos-7.md).
154
-
155
-
156
-
Make sure you have the required libraries installed by running the following shell commands:
0 commit comments