Skip to content

Commit 50e73f7

Browse files
committed
A bit more feedback
1 parent 7f68464 commit 50e73f7

File tree

1 file changed

+11
-79
lines changed

1 file changed

+11
-79
lines changed

articles/cognitive-services/Speech-Service/speech-sdk.md

Lines changed: 11 additions & 79 deletions
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ ms.author: dapine
1414

1515
# About the Speech SDK
1616

17-
The Speech software development kit (SDK) exposes many of the Speech service capabilities, to empower you to develop speech-enabled applications. The Speech SDK is available in many programming languages, all of which work cross-platform, except for Objective-C, which is only available on iOS and macOS.
17+
The Speech software development kit (SDK) exposes many of the Speech service capabilities, to empower you to develop speech-enabled applications. The Speech SDK is available in many programming languages and across all platforms.
1818

1919
[!INCLUDE [Speech SDK Platforms](../../../includes/cognitive-services-speech-service-speech-sdk-platforms.md)]
2020

@@ -56,21 +56,24 @@ The Speech SDK can be used for transcribing call center scenarios, where telepho
5656

5757
#### Call Center Transcription
5858

59-
A common scenario for speech-to-text is transcribing large volumes of telephony data that may come from various systems, such as Interactive Voice Response (IVR). The latest speech recognition models from the Speech service excel at transcribing this telephony data, even in cases when the data is difficult for a human to understand.
59+
[Call Center Transcription](call-center-transcription.md) is common scenario for speech-to-text for transcribing large volumes of telephony data that may come from various systems, such as Interactive Voice Response (IVR). The latest speech recognition models from the Speech service excel at transcribing this telephony data, even in cases when the data is difficult for a human to understand.
6060

6161
### Codec compressed audio input
6262

6363
Several of the Speech SDK programming languages support codec compressed audio input streams. For more information, see <a href="https://docs.microsoft.com/azure/cognitive-services/speech-service/how-to-use-codec-compressed-audio-input-streams" target="_blank">use compressed audio input formats <span class="docon docon-navigate-external x-hidden-focus"></span></a>.
6464

6565
## REST API
6666

67-
While the Speech SDK covers many feature capabilities of the Speech Service, for some scenarios you might want to use the REST API. Certain functionalities are only available from the Azure portal, Custom Speech portal, Custom Voice portal, or the REST API. As an example, endpoint management is only exposed via the REST API.
67+
While the Speech SDK covers many feature capabilities of the Speech Service, for some scenarios you might want to use the REST API. As an example, endpoint management is only exposed via the REST API.
6868

6969
> [!TIP]
70-
> When relying on the REST API, use the <a href="https://editor.swagger.io/" target="_blank">Swagger Editor <span class="docon docon-navigate-external x-hidden-focus"></span></a> to automatically generate client libraries.
71-
> For example, to generate a Batch transcription client library:
72-
> 1. Select **File** > **Import URL**
73-
> 1. Paste `https://westus.cris.ai/swagger/ui/index#/Custom%20Speech%20transcriptions%3A`
70+
> When relying on the REST API, use the Swagger Editor to automatically generate client libraries. For example, to generate a Batch transcription client library.
71+
> 1. Copy the example URL below:
72+
> ```http
73+
> https://westus.cris.ai/docs/v2.0/swagger
74+
> ```
75+
> 1. Navigate to the <a href="https://editor.swagger.io/" target="_blank">Swagger Editor <span class="docon docon-navigate-external x-hidden-focus"></span></a>
76+
> 1. Select **File** > **Import URL** and paste the URL
7477
> 1. Select **Generate Client** and choose your desired programming language
7578
7679
### Batch transcription
@@ -89,78 +92,7 @@ When using speech-to-text for recognition and transcription in a unique environm
8992
9093
Custom text-to-speech, also known as Custom Voice is a set of online tools that allow you to create a recognizable, one-of-a-kind voice for your brand. The creation and management of no-code Custom Voice models is available through the [Custom Voice Portal](https://aka.ms/customvoice). Once the Custom Voice model is published, it can be consumed by the Speech SDK.
9194
92-
## Get the SDK
93-
94-
# [Windows](#tab/windows)
95-
96-
> [!WARNING]
97-
> The Speech SDK supports Windows 10 and Windows Server 2016, or later versions. Earlier versions are **not supported**.
98-
99-
The Speech SDK requires the <a href="https://support.microsoft.com/help/2977003/the-latest-supported-visual-c-downloads" target="_blank">Microsoft Visual C++ Redistributable for Visual Studio 2019 <span class="docon docon-navigate-external x-hidden-focus"></span></a> on the system.
100-
101-
- <a href="https://aka.ms/vs/16/release/vc_redist.x86.exe" target="_blank">Install for x86 <span class="docon docon-navigate-external x-hidden-focus"></span></a>
102-
- <a href="https://aka.ms/vs/16/release/vc_redist.x64.exe" target="_blank">Install for x64 <span class="docon docon-navigate-external x-hidden-focus"></span></a>
103-
- <a href="https://aka.ms/vs/16/release/vc_redist.arm64.exe" target="_blank">Install for ARMx64 <span class="docon docon-navigate-external x-hidden-focus"></span></a>
104-
105-
For microphone input, the Media Foundation libraries must be installed. These libraries are part of Windows 10 and Windows Server 2016. It's possible to use the Speech SDK without these libraries, as long as a microphone isn't used as the audio input device.
106-
107-
The required Speech SDK files can be deployed in the same directory as your application. This way your application can directly access the libraries. Make sure you select the correct version (x86/x64) that matches your application.
108-
109-
| Name | Function |
110-
|-------------------------------------------------|------------------------------------------------------|
111-
| `Microsoft.CognitiveServices.Speech.core.dll` | Core SDK, required for native and managed deployment |
112-
| `Microsoft.CognitiveServices.Speech.csharp.dll` | Required for managed deployment |
113-
114-
> [!NOTE]
115-
> Starting with the release 1.3.0 the file `Microsoft.CognitiveServices.Speech.csharp.bindings.dll` (shipped in previous releases) isn't needed anymore. The functionality is now integrated in the core SDK.
116-
117-
> [!NOTE]
118-
> For the Windows Forms App (.NET Framework) C# project, make sure the libraries are included in your project's deployment settings. You can check this under `Properties -> Publish Section`. Click the `Application Files` button and find corresponding libraries from the scroll down list. Make sure the value is set to `Included`. Visual Studio will include the file when project is published/deployed.
119-
120-
For Windows, we support the following languages:
121-
122-
* C# (UWP and .NET), C++:
123-
You can reference and use the latest version of our Speech SDK NuGet package. The package includes 32-bit and 64-bit client libraries and managed (.NET) libraries. The SDK can be installed in Visual Studio by using NuGet, [Microsoft.CognitiveServices.Speech](https://www.nuget.org/packages/Microsoft.CognitiveServices.Speech).
124-
125-
* Java:
126-
You can reference and use the latest version of our Speech SDK Maven package, which supports only Windows x64. In your Maven project, add `https://csspeechstorage.blob.core.windows.net/maven/` as an additional repository and reference `com.microsoft.cognitiveservices.speech:client-sdk:1.8.0` as a dependency.
127-
128-
# [Linux](#tab/linux)
129-
130-
> [!NOTE]
131-
> Currently, we only support the following distributions and development languages/platforms:
132-
>
133-
> | Distribution | Development |
134-
> |:-|:-|
135-
> |Ubuntu 16.04 x86 |C++|
136-
> |Ubuntu 16.04 x64 |C++, Java, .NET Core, Python|
137-
> |Ubuntu 16.04 ARM32 |C++, Java, .NET Core|
138-
> |Ubuntu 16.04 ARM64 |C++, Java, .NET Core[<sup>[1]</sup>](#footnote1)|
139-
> |Ubuntu 18.04 x86 |C++|
140-
> |Ubuntu 18.04 x64 |C++, Java, .NET Core, Python|
141-
> |Ubuntu 18.04 ARM32 |C++, Java, .NET Core|
142-
> |Ubuntu 18.04 ARM64 |C++, Java, .NET Core[<sup>[1]</sup>](#footnote1)|
143-
> |Debian 9 x86 |C++|
144-
> |Debian 9 x64 |C++, Java, .NET Core, Python|
145-
> |Debian 9 ARM32 |C++, Java, .NET Core|
146-
> |Debian 9 ARM64 |C++, Java, .NET Core[<sup>[1]</sup>](#footnote1)|
147-
> |Red Hat Enterprise Linux (RHEL) 7 x64[<sup>[2]</sup>](#footnote2) |C++, Java, .NET Core, Python|
148-
> |Red Hat Enterprise Linux (RHEL) 8 x64 |C++, Java, .NET Core, Python|
149-
> |CentOS 7 x64[<sup>[2]</sup>](#footnote2) |C++, Java, .NET Core, Python|
150-
> |CentOS 8 x64 |C++, Java, .NET Core, Python|
151-
>
152-
> **[<a name="footnote1">1</a>]** Linux ARM64 requires .NET Core 3.x (dotnet-sdk-3.x package) for proper ARM64 support.<br>
153-
> **[<a name="footnote2">2</a>]** Follow the instructions on [how to configure RHEL/CentOS 7 for Speech SDK](~/articles/cognitive-services/speech-service/how-to-configure-rhel-centos-7.md).
154-
155-
156-
Make sure you have the required libraries installed by running the following shell commands:
157-
158-
On Ubuntu:
159-
160-
```sh
161-
sudo apt-get update
162-
sudo apt-get install libssl1.0.0 libasound2
163-
```
95+
## Get the Speech SDK
16496
16597
# [Windows](#tab/windows)
16698

0 commit comments

Comments
 (0)