Skip to content

Commit 0a52e19

Browse files
authored
Merge pull request #229322 from BrianMouncer/BrianMouncer-pr/1.26.0_updates
Brian mouncer pr/1.26.0 updates
2 parents 3293136 + 0e6bae6 commit 0a52e19

File tree

10 files changed

+45
-42
lines changed

10 files changed

+45
-42
lines changed

articles/cognitive-services/Speech-Service/embedded-speech.md

Lines changed: 9 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -58,7 +58,7 @@ The Speech SDK for Java doesn't support Windows on ARM64.
5858

5959
Embedded speech is only available with C#, C++, and Java SDKs. The other Speech SDKs, Speech CLI, and REST APIs don't support embedded speech.
6060

61-
Embedded speech recognition only supports mono 16 bit, 16-kHz PCM-encoded WAV audio.
61+
Embedded speech recognition only supports mono 16 bit, 8-kHz or 16-kHz PCM-encoded WAV audio.
6262

6363
Embedded neural voices only support 24-kHz sample rate.
6464

@@ -114,10 +114,10 @@ embeddedSpeechConfig.SetSpeechSynthesisVoice(
114114
embeddedSpeechConfig.SetSpeechSynthesisOutputFormat(SpeechSynthesisOutputFormat.Riff24Khz16BitMonoPcm);
115115
```
116116

117-
You can find ready to use embedded speech samples at [GitHub](https://aka.ms/csspeech/samples).
117+
You can find ready to use embedded speech samples at [GitHub](https://aka.ms/embedded-speech-samples). For remarks on projects from scratch, see samples specific documentation:
118118

119-
- [C# (.NET 6.0)](https://github.com/Azure-Samples/cognitive-services-speech-sdk/tree/master/samples/csharp/dotnetcore/embedded-speech)
120-
- [C# for Unity](https://github.com/Azure-Samples/cognitive-services-speech-sdk/tree/master/samples/csharp/unity/embedded-speech)
119+
- [C# (.NET 6.0)](https://aka.ms/embedded-speech-samples-csharp)
120+
- [C# for Unity](https://aka.ms/embedded-speech-samples-csharp-unity)
121121
::: zone-end
122122

123123
::: zone pivot="programming-language-cpp"
@@ -144,7 +144,8 @@ embeddedSpeechConfig->SetSpeechSynthesisVoice(
144144
embeddedSpeechConfig->SetSpeechSynthesisOutputFormat(SpeechSynthesisOutputFormat::Riff24Khz16BitMonoPcm);
145145
```
146146
147-
You can find ready to use embedded speech samples at [GitHub](https://github.com/Azure-Samples/cognitive-services-speech-sdk/tree/master/samples/cpp/embedded-speech)
147+
You can find ready to use embedded speech samples at [GitHub](https://aka.ms/embedded-speech-samples). For remarks on projects from scratch, see samples specific documentation:
148+
- [C++](https://aka.ms/embedded-speech-samples-cpp)
148149
::: zone-end
149150
150151
::: zone pivot="programming-language-java"
@@ -168,9 +169,9 @@ embeddedSpeechConfig.setSpeechSynthesisVoice(
168169
embeddedSpeechConfig.setSpeechSynthesisOutputFormat(SpeechSynthesisOutputFormat.Riff24Khz16BitMonoPcm);
169170
```
170171

171-
You can find ready to use embedded speech samples at [GitHub](https://aka.ms/csspeech/samples).
172-
- [Java (JRE)](https://github.com/Azure-Samples/cognitive-services-speech-sdk/tree/master/samples/java/jre/embedded-speech)
173-
- [Java for Android](https://github.com/Azure-Samples/cognitive-services-speech-sdk/tree/master/samples/java/android/embedded-speech)
172+
You can find ready to use embedded speech samples at [GitHub](https://aka.ms/embedded-speech-samples). For remarks on projects from scratch, see samples specific documentation:
173+
- [Java (JRE)](https://aka.ms/embedded-speech-samples-java)
174+
- [Java for Android](https://aka.ms/embedded-speech-samples-java-android)
174175
::: zone-end
175176

176177

articles/cognitive-services/Speech-Service/includes/pattern-matching-overview.md

Lines changed: 22 additions & 20 deletions
Original file line numberDiff line numberDiff line change
@@ -10,15 +10,15 @@ ms.author: chschrae
1010
keywords: intent recognition pattern matching
1111
---
1212

13-
Pattern matching can be customized to group together pattern intents and entities inside a ``PatternMatchingModel``. Using this grouping, it's possible to access more advanced entity types that will help make your intent recognition more precise.
13+
Pattern matching can be customized to group together pattern intents and entities inside a ``PatternMatchingModel``. Using this grouping, it's possible to access more advanced entity types that help make your intent recognition more precise.
1414

15-
For supported locales see [here](../language-support.md?tabs=intent-recognition).
15+
For supported locales, see [here](../language-support.md?tabs=intent-recognition).
1616

1717
## Patterns vs. Exact Phrases
1818

1919
There are two types of strings used in the pattern matcher: "exact phrases" and "patterns". It's important to understand the differences.
2020

21-
Exact phrases are a string of the exact words that you'll want to match. For example:
21+
Exact phrases are a strings of the exact words that you want to match. For example:
2222

2323
> "Take me to floor seven".
2424
@@ -32,58 +32,58 @@ The ``PatternMatchingModel`` contains an ID to reference that model by, a list o
3232

3333
### Pattern Matching Intents
3434

35-
``PatternMatchingIntent`` objects represent a collection of phrases that will be used to evaluate speech or text in the ``IntentRecognizer``. If the phrases are matched, the ``IntentRecognitionResult`` returned will have the ID of the ``PatternMatchingIntent`` that was matched.
35+
``PatternMatchingIntent`` objects represent a collection of phrases that are used to evaluate speech or text in the ``IntentRecognizer``. If the phrases are matched, then the ``IntentRecognitionResult`` returned will have the ID of the ``PatternMatchingIntent`` that was matched.
3636

3737
### Pattern Matching Entities
3838

39-
``PatternMatchingEntity`` objects represent an individual entity reference and its corresponding properties that tell the ``IntentRecognizer`` how to treat it. All ``PatternMatchingEntity`` objects must have an ID that is present in a phrase or else it will never be matched.
39+
``PatternMatchingEntity`` objects represent an individual entity reference and its corresponding properties that tell the ``IntentRecognizer`` how to treat it. All ``PatternMatchingEntity`` objects must have an ID that is present in a phrase or else it will not be matched.
4040

4141
#### Entity Naming restrictions
4242

43-
Entity names containing ':' characters will assign a role to an entity. (See below)
43+
Entity names containing ':' characters assign a role to an entity. (See below)
4444

4545
## Types of Entities
4646

4747
### Any Entity
4848

49-
The "Any" entity will match any text that appears in that slot regardless of the text it contains. If we consider our previous example using the pattern "Take me to floor {floorName}", the user might say something like:
49+
The "Any" entity matches any text that appears in that slot regardless of the text it contains. If we consider our previous example using the pattern "Take me to floor {floorName}", the user might say something like:
5050

5151
> "Take me to the floor parking 2
5252
5353
In this case, the "floorName" entity would match "parking 2".
5454

55-
These are lazy matches that will attempt to match as few words as possible unless it appears at the beginning or end of an utterance. Consider the pattern:
55+
These entities are lazy matches that attempt to match as few words as possible unless it appears at the beginning or end of an utterance. Consider the following pattern:
5656

5757
> "Take me to the floor {floorName1} {floorName2}"
5858
5959
In this case, the utterance "Take me to the floor parking 2" would match and return floorName1 = "parking" and floorName2 = "2".
6060

61-
It may be tricky to handle extra text if it's captured. Perhaps the user kept talking and the utterance captured more than their command. "Take me to floor parking 2 yes Janice I heard about that let's". In this case the floorName1 would be correct, but floorName2 would = "2 yes Janice I heard about that let's". It's important to be aware of the way the Entities will match and adjust to your scenario appropriately. The Any entity type is the most basic and least precise.
61+
It may be tricky to handle extra text if it's captured. Perhaps the user kept talking and the utterance captured more than their command. "Take me to floor parking 2 yes Janice I heard about that let's". In this case the floorName1 would be correct, but floorName2 would = "2 yes Janice I heard about that let's". It's important to be aware of the way the Entities match, and adjust your scenario appropriately. The Any entity type is the most basic and least precise type of matching done.
6262

6363
### List Entity
6464

65-
The "List" entity is made up of a list of phrases that will guide the engine on how to match it. The "List" entity has two modes. "Strict" and "Fuzzy".
65+
The "List" entity is made up of a list of phrases that guide the engine on how to match it. The "List" entity has two modes. "Strict" and "Fuzzy".
6666

67-
Let's assume we have a list of floors for our elevator. Since we're dealing with speech, we will add entries for the lexical format as well.
67+
Let's assume we have a list of floors for our elevator. Since we're dealing with speech, we add entries using the lexical format as well.
6868

6969
> "1", "2", "3", "lobby", "ground floor", "one", "two", "three"
7070
71-
When an entity with an ID is of type "List" and is in "Strict" mode, the engine will only match if the text in the slot appears in the list.
71+
When an entity of type ID "List" is used in "Strict" mode, the engine only matches if the text in the slot appears in the list.
7272

7373
> "take me to floor one" will match.
7474
7575
> "take me to floor 5" will not.
7676
77-
It's important to note that the Intent will not match, not just the entity.
77+
It's important to note that the entire Intent will not match, not just the entity.
7878

79-
When an entity is of type "List" and is in "Fuzzy" mode, the engine will still match the Intent, and will return the text that appeared in the slot in the utterance even if it's not in the list. This is useful behind the scenes to help make the speech recognition better.
79+
When an entity of type ID "List" is used in "Fuzzy" mode, the engine still matches the Intent, and will return the text that appeared in the slot in the utterance, even if it's not in the list. This is useful behind the scenes to help make the speech recognition better.
8080

8181
> [!WARNING]
8282
> Fuzzy list entities are implemented, but not integrated into the speech recognition part. Therefore, they will match entities, but not improve speech recognition.
8383
8484
### Prebuilt Integer Entity
8585

86-
The "PrebuiltInteger" entity is used when you expect to get an integer in that slot. It won't match the intent if an integer cannot be found. The return value is a string representation of the number.
86+
The "PrebuiltInteger" entity is used when you expect to get an integer in that slot. It will not match the intent if an integer cannot be found. The return value is a string representation of the number.
8787

8888
### Examples of a valid match and return values
8989

@@ -111,13 +111,13 @@ Consider our elevator example.
111111

112112
> "Take me to floor {floorName}"
113113
114-
If "floorName" is a prebuilt integer entity, the expectation is that whatever text is inside the slot will represent an integer. Here a floor number would match well, but a floor with a name such as "lobby" would not.
114+
If "floorName" is a prebuilt integer entity, the expectation is that whatever text is inside the slot represents an integer. Here a floor number would match well, but a floor with a name such as "lobby" would not.
115115

116116
## Grouping required and optional items
117117

118-
In the pattern it's allowed to include words or entities that may be present in the utterance or not. This is especially useful for determiners like "the", "a", or "an". This doesn't have any functional difference from hard coding out the many combinations, but can help reduce the number of patterns needed. Indicate optional items with "[" and "]". Indicate required items with "(" and ")". You may include multiple items in the same group by separating them with a '|' character.
118+
In the pattern, it is allowed to include words or entities that "might" be present in the utterance. This is especially useful for determiners like "the", "a", or "an". This doesn't have any functional difference from hard coding out the many combinations, but can help reduce the number of patterns needed. Indicate optional items with "[" and "]". Indicate required items with "(" and ")". You may include multiple items in the same group by separating them with a '|' character.
119119

120-
To see how this would reduce the number of patterns needed consider the following set.
120+
To see how this would reduce the number of patterns needed, consider the following set:
121121

122122
> "Take me to {floorName}"
123123
@@ -143,7 +143,7 @@ It's also possible to include optional entities. Imagine there are multiple park
143143

144144
>"Take me to [{floorType}] {floorName}"
145145
146-
Optionals are also very useful if you might be using keyword recognition and a push-to-talk function. This means sometimes the keyword will be present, and sometimes it won't. Assuming your keyword was "computer" your pattern would look something like this.
146+
Optionals are also useful if you might be using keyword recognition and a push-to-talk function. This means sometimes the keyword will be present, and sometimes it won't. Assuming your keyword was "computer" your pattern would look something like this.
147147

148148
>"[Computer] Take me to {floorName}"
149149
@@ -160,13 +160,15 @@ Given a pattern like this, there will be two entities in the result labeled "cit
160160

161161
## Intent Matching Priority
162162

163-
Sometimes multiple patterns may match the same utterance. In this case, the engine will give priority to patterns as follows.
163+
Sometimes multiple patterns may match the same utterance. In this case, the engine gives priority to patterns as follows.
164164

165165
1. Exact Phrases.
166166
2. Patterns with more Entities.
167167
3. Patterns with Integer Entities.
168168
4. Patterns with List Entities.
169169
5. Patterns with Any Entities.
170+
6. Patterns with more bytes matched.
171+
- Example: Pattern "click {something} on the left" will be higher priority than "click {something}".
170172

171173
## Next steps
172174

articles/cognitive-services/Speech-Service/includes/quickstarts/captioning/java.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@ Before you can do anything, you need to [install the Speech SDK](~/articles/cogn
4343
<dependency>
4444
<groupId>com.microsoft.cognitiveservices.speech</groupId>
4545
<artifactId>client-sdk</artifactId>
46-
<version>1.24.2</version>
46+
<version>1.26.0</version>
4747
</dependency>
4848
</dependencies>
4949
</project>

articles/cognitive-services/Speech-Service/includes/quickstarts/platform/java-android.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -48,6 +48,6 @@ Add the Speech SDK as a dependency in your project.
4848
1. Select **File** > **Project structure** > **Dependencies** > **app**.
4949
1. Select the plus symbol (**+**) to add a dependency under **Declared Dependencies**. Then select **Library dependency** from the drop-down menu.
5050
:::image type="content" source="../../../media/sdk/android-studio/sdk-install-3-zoom.png" alt-text="Screenshot that shows how to add a library dependency in Android Studio." lightbox="../../../media/sdk/android-studio/sdk-install-3.png":::
51-
1. In the **Add Library Dependency** window that appears, enter the name and version of the Speech SDK for Java: **com.microsoft.cognitiveservices.speech:client-sdk:1.24.2**. Then select **Search**.
51+
1. In the **Add Library Dependency** window that appears, enter the name and version of the Speech SDK for Java: **com.microsoft.cognitiveservices.speech:client-sdk:1.26.0**. Then select **Search**.
5252
1. Make sure that the selected **Group ID** is **com.microsoft.cognitiveservices.speech**, and then select **OK**.
5353
1. Select **OK** to close the **Project Structure** window and apply your changes to the project.

articles/cognitive-services/Speech-Service/includes/quickstarts/platform/java-jre.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -54,7 +54,7 @@ Follow these steps to install the Speech SDK for Java using Apache Maven:
5454
<dependency>
5555
<groupId>com.microsoft.cognitiveservices.speech</groupId>
5656
<artifactId>client-sdk</artifactId>
57-
<version>1.24.2</version>
57+
<version>1.26.0</version>
5858
</dependency>
5959
</dependencies>
6060
</project>
@@ -107,7 +107,7 @@ Follow these steps to install the Speech SDK for Java using Apache Maven:
107107
<dependency>
108108
<groupId>com.microsoft.cognitiveservices.speech</groupId>
109109
<artifactId>client-sdk</artifactId>
110-
<version>1.24.2</version>
110+
<version>1.26.0</version>
111111
</dependency>
112112
</dependencies>
113113
```
@@ -125,7 +125,7 @@ Gradle configurations require an explicit reference to the .jar dependency exten
125125
// build.gradle
126126

127127
dependencies {
128-
implementation group: 'com.microsoft.cognitiveservices.speech', name: 'client-sdk', version: "1.24.2", ext: "jar"
128+
implementation group: 'com.microsoft.cognitiveservices.speech', name: 'client-sdk', version: "1.26.0", ext: "jar"
129129
}
130130
```
131131
---

articles/cognitive-services/Speech-Service/includes/quickstarts/platform/objectivec.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@ The macOS CocoaPod package is available for download and use with the [Xcode 9.4
3434
use_frameworks!
3535
3636
target 'AppName' do
37-
pod 'MicrosoftCognitiveServicesSpeech-macOS', '~> 1.25.0'
37+
pod 'MicrosoftCognitiveServicesSpeech-macOS', '~> 1.26.0'
3838
end
3939
```
4040
1. Run `pod install` to install the Speech SDK.
@@ -64,12 +64,12 @@ The macOS CocoaPod package is available for download and use with the [Xcode 9.4
6464
use_frameworks!
6565
6666
target 'AppName' do
67-
pod 'MicrosoftCognitiveServicesSpeech-iOS', '~> 1.25.0'
67+
pod 'MicrosoftCognitiveServicesSpeech-iOS', '~> 1.26.0'
6868
end
6969
```
7070
1. Run `pod install` to install the Speech SDK.
7171
72-
Alternatively, you can download the [binary CocoaPod](https://aka.ms/csspeech/iosbinary) and unzip to extract it's contents. In your Xcode project, add a reference to the extracted "MicrosoftCognitiveServicesSpeech.xcframework" folder and it's contents.
72+
Alternatively, you can download the [binary CocoaPod](https://aka.ms/csspeech/iosbinary) and unzip to extract its contents. In your Xcode project, add a reference to the extracted "MicrosoftCognitiveServicesSpeech.xcframework" folder and it's contents.
7373
7474
> [!NOTE]
7575
> .NET developers can build native iOS applications by using the Xamarin.iOS application framework For more information, see [Xamarin.iOS](/xamarin/ios/).

articles/cognitive-services/Speech-Service/includes/quickstarts/platform/swift.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@ The macOS CocoaPod package is available for download and use with the [Xcode 9.4
3434
use_frameworks!
3535
3636
target 'AppName' do
37-
pod 'MicrosoftCognitiveServicesSpeech-macOS', '~> 1.25.0'
37+
pod 'MicrosoftCognitiveServicesSpeech-macOS', '~> 1.26.0'
3838
end
3939
```
4040
1. Run `pod install` to install the Speech SDK.
@@ -65,12 +65,12 @@ The macOS CocoaPod package is available for download and use with the [Xcode 9.4
6565
use_frameworks!
6666
6767
target 'AppName' do
68-
pod 'MicrosoftCognitiveServicesSpeech-iOS', '~> 1.25.0'
68+
pod 'MicrosoftCognitiveServicesSpeech-iOS', '~> 1.26.0'
6969
end
7070
```
7171
1. Run `pod install` to install the Speech SDK.
7272
73-
Alternatively, you can download the [binary CocoaPod](https://aka.ms/csspeech/iosbinary) and unzip to extract it's contents. In your Xcode project, add a reference to the extracted "MicrosoftCognitiveServicesSpeech.xcframework" folder and it's contents.
73+
Alternatively, you can download the [binary CocoaPod](https://aka.ms/csspeech/iosbinary) and unzip to extract its contents. In your Xcode project, add a reference to the extracted "MicrosoftCognitiveServicesSpeech.xcframework" folder and it's contents.
7474
7575
> [!NOTE]
7676
> .NET developers can build native iOS applications by using the Xamarin.iOS application framework For more information, see [Xamarin.iOS](/xamarin/ios/).

articles/cognitive-services/Speech-Service/includes/quickstarts/speech-to-text-basics/java.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@ Before you can do anything, you need to install the Speech SDK. The sample in th
4343
<dependency>
4444
<groupId>com.microsoft.cognitiveservices.speech</groupId>
4545
<artifactId>client-sdk</artifactId>
46-
<version>1.24.2</version>
46+
<version>1.26.0</version>
4747
</dependency>
4848
</dependencies>
4949
</project>

articles/cognitive-services/Speech-Service/includes/quickstarts/speech-translation-basics/java.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@ Before you can do anything, you need to install the Speech SDK. The sample in th
4343
<dependency>
4444
<groupId>com.microsoft.cognitiveservices.speech</groupId>
4545
<artifactId>client-sdk</artifactId>
46-
<version>1.24.2</version>
46+
<version>1.26.0</version>
4747
</dependency>
4848
</dependencies>
4949
</project>

articles/cognitive-services/Speech-Service/includes/quickstarts/text-to-speech-basics/java.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@ Before you can do anything, you need to install the Speech SDK. The sample in th
4343
<dependency>
4444
<groupId>com.microsoft.cognitiveservices.speech</groupId>
4545
<artifactId>client-sdk</artifactId>
46-
<version>1.24.2</version>
46+
<version>1.26.0</version>
4747
</dependency>
4848
</dependencies>
4949
</project>

0 commit comments

Comments
 (0)