You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/communication-services/concepts/call-automation/azure-communication-services-azure-cognitive-services-integration.md
# Connect Azure Communication Services with Azure AI services
16
16
17
-
Azure Communication Services Call Automation APIs provide developers the ability to steer and control the Azure Communication Services Telephony, VoIP or WebRTC calls using real-time event triggers to perform actions based on custom business logic specific to their domain. Within the Call Automation APIs developers can use simple AIpowered APIs, which can be used to play personalized greeting messages, recognize conversational voice inputs to gather information on contextual questions to drive a more self-service model with customers, use sentiment analysis to improve customer service overall. These content specific APIs are orchestrated through **Azure AI Services** with support for customization of AI models without developers needing to terminate media streams on their services and streaming back to Azure for AI functionality.
17
+
Azure Communication Services Call Automation APIs allow developers to steer and control calls made through Azure Communication Services, including telephony, VoIP, and WebRTC. These APIs use real-time event triggers, which enable actions based on custom business logic that is specific to each developer’s domain. With Call Automation APIs, developers can use simple AI-powered features. For example, they can play personalized greetings, recognize spoken responses to gather information from customers, and analyze sentiment to improve service. These targeted APIs are managed through **Azure AI Foundry**, which allow developers to customize AI models. Importantly, developers do not need to deal with media streams or send them back to Azure for these AI functions—the processing happens seamlessly.
18
18
19
-
All this is possible with one-click where enterprises can access a secure solution and link their models through the portal. Furthermore, developers and enterprises don't need to manage credentials. Connecting your Azure AI services uses managed identities to access user-owned resources. Developers can use managed identities to authenticate any resource that supports Microsoft Entra authentication.
19
+
All this functionality is possible with oneclick, allowing enterprises to access a secure solution and link their models through the portal. Furthermore, developers and enterprises don't need to manage credentials. Connecting your Azure AI services uses managed identities to access user-owned resources. Developers can use managed identities to authenticate any resource that supports Microsoft Entra authentication.
20
20
21
21
Azure AI services can be easily integrated into any application regardless of the programming language. When creating an Azure Resource in Azure portal, enable the option and provide the URL to the Azure AI services. This simple experience allows developers to meet their needs, scale, and avoid investing time and resources into designing and maintaining a custom solution.
22
22
23
23
> [!NOTE]
24
-
> This integration is supported in limited regions for Azure AI services, for more information about which regions are supported please view the limitations section at the bottom of this document. This integration only supports Multi-service Cognitive Service resource, we recommend if you're creating a new Azure AI Service resource you create a Multi-service Cognitive Service resource or when you're connecting an existing resource confirm that it is a Multi-service Cognitive Service resource.
24
+
> This integration only supports Multi-service Cognitive Service resource, we recommend if you're creating a new Azure AI Service resource you create a Multi-service Cognitive Service resource or when you're connecting an existing resource confirm that it's a Multi-service Cognitive Service resource.
25
25
26
26
## Common use cases
27
27
28
28
### Build applications that can play and recognize speech
29
29
30
-
With the ability to connect your Azure AI services to Azure Communication Services. You can enable custom play functionality, using [Text-to-Speech](../../../../articles/cognitive-services/Speech-Service/text-to-speech.md) and [Speech Synthesis Markup Language (SSML)](../../../../articles/cognitive-services/Speech-Service/speech-synthesis-markup.md) configuration, to play more customized and natural sounding audio to users. Through the Azure AI services connection, you can also use the Speech-To-Text service to incorporate recognition of voice responses that can be converted into actionable tasks through business logic in the application. These functions can be further enhanced through the ability to create custom models within Azure AI services that are bespoke to your domain and region, through the ability to choose which languages are spoken and recognized, custom voices and custom models built based on your experience.
30
+
With the ability to connect your Azure AI services to Azure Communication Services. You can enable custom play functionality, using [Text-to-Speech](../../../../articles/cognitive-services/Speech-Service/text-to-speech.md) and [Speech Synthesis Markup Language (SSML)](../../../../articles/cognitive-services/Speech-Service/speech-synthesis-markup.md) configuration, to play more customized and natural sounding audio to users. Through the Azure AI services connection, you can also use the Speech-To-Text service to incorporate recognition of voice responses that can be converted into actionable tasks through business logic in the application. These functions can be further enhanced within Azure AI services by:
31
+
32
+
- Creating custom models tailored to your domain and region
33
+
- Selecting which languages are spoken and recognized
34
+
- Designing custom voices
35
+
- Building additional models based on your experience
31
36
32
37
## Runtime flow
33
38
[](./media/run-time-flow.png#lightbox)
34
39
35
40
## Azure portal experience
36
-
You'll need to connect your Azure Communication Services resource with the Azure AI resource through the Azure portal. There are two ways you can accomplish this step:
37
-
-By navigating through the steps of the Cognitive Services tab in your Azure Communication Services (recommended).
41
+
You need to connect your Azure Communication Services resource with the Azure AI resource through the Azure portal. There are two ways you can accomplish this step:
42
+
-Navigating through the steps of the Cognitive Services tab in your Azure Communication Services (recommended).
38
43
- Manually adding the Managed Identity to your Azure Communication Services resource. This step is more advanced and requires a little more effort to connect your Azure Communication Services to your Azure AI services.
39
44
40
45
## Prerequisites
@@ -46,7 +51,7 @@ You'll need to connect your Azure Communication Services resource with the Azure
46
51
### Connecting through the Azure portal
47
52
48
53
1. Open your Azure Communication Services resource and click on the Cognitive Services tab.
49
-
2. If system-assigned managed identity isn't enabled, you'll need to enable it.
54
+
2. If system-assigned managed identity isn't enabled, you need to enable it.
50
55
3. In the Cognitive Services tab, click on "Enable Managed Identity" button.
51
56
52
57
[](./media/enabled-identity.png#lightbox)
@@ -57,7 +62,7 @@ You'll need to connect your Azure Communication Services resource with the Azure
57
62
5. Once the identity is enabled, you should see something similar.
58
63
[](./media/identity-saved.png#lightbox)
59
64
60
-
6. When managed identity is enabled the Cognitive Service tab should show a button 'Connect cognitive service' to connect the two services.
65
+
6. When managed identity is enabled, the Cognitive Service tab should show a button 'Connect cognitive service' to connect the two services.
61
66
[](./media/cog-svc.png#lightbox)
62
67
63
68
7. Click on 'Connect cognitive service', select the Subscription, Resource Group and Resource and click 'Connect' in the context pane that opens up.
@@ -109,27 +114,9 @@ Your Azure Communication Service has now been linked to your Azure Cognitive Ser
109
114
110
115
## Azure AI services regions supported
111
116
112
-
This integration between Azure Communication Services and Azure AI services is only supported in the following regions:
113
-
- centralus
114
-
- northcentralus
115
-
- southcentralus
116
-
- westcentralus
117
-
- eastus
118
-
- eastus2
119
-
- westus
120
-
- westus2
121
-
- westus3
122
-
- canadacentral
123
-
- northeurope
124
-
- westeurope
125
-
- uksouth
126
-
- southafricanorth
127
-
- centralindia
128
-
- eastasia
129
-
- southeastasia
130
-
- australiaeast
131
-
- brazilsouth
132
-
- uaenorth
117
+
Our integration between Azure Communication Services (ACS) and Azure AI is fully aligned with the regional availability of Azure AI Foundry. This means that ACS to Azure AI integration is supported in all regions where Azure AI Foundry is available.
118
+
119
+
For the most up-to-date list of supported regions, refer to the official Azure AI Foundry [documentation](/azure/ai-services/speech-service/regions).
Copy file name to clipboardExpand all lines: articles/communication-services/concepts/call-automation/play-action.md
+7-6Lines changed: 7 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -12,15 +12,15 @@ ms.author: kpunjabi
12
12
# Playing audio in call
13
13
14
14
The play action provided through the Azure Communication Services Call Automation SDK allows you to play audio prompts to participants in the call. This action can be accessed through the server-side implementation of your application. You can play audio to call participants through one of two methods;
15
-
- Providing Azure Communication Services access to prerecorded audio files of WAV format, that Azure Communication Services can access with support for authentication
15
+
- Providing Azure Communication Services access to prerecorded audio files in WAV format, which Azure Communication Services can access with support for authentication.
16
16
- Regular text that can be converted into speech output through the integration with Azure AI services.
17
17
18
18
You can use the newly announced integration between [Azure Communication Services and Azure AI services](./azure-communication-services-azure-cognitive-services-integration.md) to play personalized responses using Azure [Text-To-Speech](../../../../articles/cognitive-services/Speech-Service/text-to-speech.md). You can use human like prebuilt neural voices out of the box or create custom neural voices that are unique to your product or brand. For more information on supported voices, languages and locales see [Language and voice support for the Speech service](../../../../articles/cognitive-services/Speech-Service/language-support.md).
19
19
> [!NOTE]
20
-
> Azure Communication Services currently supports two file formats, MP3 files with ID3V2TAG and WAV files formatted as 16-bit PCM mono channel audio recorded at 16KHz. You can create your own audio files using [Speech synthesis with Audio Content Creation tool](/azure/ai-services/speech-service/how-to-audio-content-creation).
20
+
> Azure Communication Services currently supports two file formats, MP3 files with ID3V2TAG and WAV files formatted as 16-bit Pulse Code Modulation (PCM) mono channel audio recorded at 16 KHz. You can create your own audio files using [Speech synthesis with Audio Content Creation tool](/azure/ai-services/speech-service/how-to-audio-content-creation).
21
21
22
22
## Prebuilt Neural Text to Speech voices
23
-
Microsoft uses deep neural networks to overcome the limits of traditional speech synthesis with regard to stress and intonation in spoken language. Prosody prediction and voice synthesis occur simultaneously, resulting in a more fluid and natural sounding output. You can use these neural voices to make interactions with your chatbots and voice assistants more natural and engaging. There are over 100 prebuilt voices to choose from. Learn more about [Azure Text-to-Speech voices](../../../../articles/cognitive-services/Speech-Service/language-support.md).
23
+
Microsoft uses advanced machine learning to overcome the limitations of traditional speech synthesis in handling stress and intonation in spoken language. Prosody prediction and voice synthesis occur simultaneously, resulting in a more fluid and natural sounding output. You can use these neural voices to make interactions with your chatbots and voice assistants more natural and engaging. There are over 100 prebuilt voices to choose from. Learn more about [Azure Text-to-Speech voices](../../../../articles/cognitive-services/Speech-Service/language-support.md).
24
24
25
25
## Common use cases
26
26
@@ -31,13 +31,13 @@ Your application might want to play some sort of announcement when a participant
31
31
32
32
### Self-serve customers
33
33
34
-
In scenarios with IVRs and virtual assistants, you can use your application or bots to play audio prompts to callers, this prompt can be in the form of a menu to guide the caller through their interaction.
34
+
In scenarios with Interactive Voice Response (IVR) systems and virtual assistants, you can use your application or bots to play audio prompts to callers, this prompt can be in the form of a menu to guide the caller through their interaction.
35
35
36
36
### Hold music
37
37
The play action can also be used to play hold music for callers. This action can be set up in a loop so that the music keeps playing until an agent is available to assist the caller.
38
38
39
39
### Playing compliance messages
40
-
As part of compliance requirements in various industries, vendors are expected to play legal or compliance messages to callers, for example, “This call is recorded for quality purposes.”.
40
+
As part of compliance requirements in various industries, vendors are expected to play legal or compliance messages to callers, for example, "This call is recorded for quality purposes."
41
41
42
42
## Sample architecture for playing audio in call using Text-To-Speech
43
43
@@ -48,7 +48,8 @@ As part of compliance requirements in various industries, vendors are expected t
48
48

49
49
50
50
## Known limitations
51
-
- Text-to-Speech text prompts support a maximum of 400 characters, if your prompt is longer than this we suggest using SSML for Text-to-Speech based play actions.
51
+
- Text-to-Speech prompts support a maximum of 4,000 characters. If your prompt is longer than this limit, we suggest using Speech Synthesis Markup Language (SSML) for Text-to-Speech based play actions.
52
+
- HD Voices from Azure AI Foundry aren't officially supported. Using these voices results in timeouts.
52
53
- For scenarios where you exceed your Speech service quota limit, you can request to increase this limit by following the steps outlined [here](/azure/ai-services/speech-service/speech-services-quotas-and-limits).
Copy file name to clipboardExpand all lines: articles/communication-services/concepts/call-automation/recognize-action.md
+2-1Lines changed: 2 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -66,7 +66,8 @@ The recognize action can be used for many reasons, here are a few examples of ho
66
66
67
67
## Known limitation
68
68
- In-band DTMF is not supported, use RFC 2833 DTMF instead.
69
-
- Text-to-Speech text prompts support a maximum of 400 characters, if your prompt is longer than this we suggest using SSML for Text-to-Speech based play actions.
69
+
- Text-to-Speech text prompts support a maximum of 4,000 characters, if your prompt is longer than this we suggest using SSML for Text-to-Speech based play actions.
70
+
- When recording is enabled speech input is recorded for 1:1 calls, but isn't recorded in group calls.
70
71
- For scenarios where you exceed your Speech service quota limit, you can request to increase this limit by following the steps outlined [here](/azure/ai-services/speech-service/speech-services-quotas-and-limits).
# Customize voice prompts to users with Play action
16
16
17
-
This guide will help you get started with playing audio files to participants by using the play action provided through Azure Communication Services Call Automation SDK.
17
+
This guide helps you get started with playing audio files to participants by using the play action provided through Azure Communication Services Call Automation SDK.
18
18
19
19
::: zone pivot="programming-language-csharp"
20
20
[!INCLUDE [Play audio with .NET](./includes/play-audio-quickstart-csharp.md)]
@@ -48,8 +48,9 @@ This guide will help you get started with playing audio files to participants by
48
48
|PlayFailed | 500 | 8572 | Action failed due to play service shutdown. |
49
49
50
50
## Known limitations
51
-
- Text-to-Speech text prompts support a maximum of 400 characters, if your prompt is longer than this we suggest using SSML for Text-to-Speech based play actions.
52
-
- For scenarios where you exceed your Speech service quota limit, you can request to increase this limit by following the steps outlined [here](/azure/ai-services/speech-service/speech-services-quotas-and-limits).
51
+
- Text-to-Speech prompts support a maximum of 4,000 characters. If your prompt is longer than this limit, we suggest using SSML for Text-to-Speech-based play actions.
52
+
- HD Voices from Azure AI Foundry aren't officially supported. These will result in timeouts if used.
53
+
- Speech service quota increases can be requested if you exceed your quota limit. Follow the steps outlined [here](/azure/ai-services/speech-service/speech-services-quotas-and-limits) to request an increase.
Copy file name to clipboardExpand all lines: articles/communication-services/how-tos/call-automation/recognize-action.md
+3-2Lines changed: 3 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -57,8 +57,9 @@ This guide helps you get started recognizing DTMF input provided by participants
57
57
58
58
## Known limitations
59
59
- In-band DTMF isn't supported. Use RFC 2833 DTMF instead.
60
-
- Text-to-Speech text prompts support a maximum of 400 characters, if your prompt is longer than this we suggest using SSML for Text-to-Speech based play actions.
61
-
- For scenarios where you exceed your Speech service quota limit, you can request to increase this limit by following the steps outlined in [Speech services quotas and limits](/azure/ai-services/speech-service/speech-services-quotas-and-limits).
60
+
- Text-to-Speech prompts support a maximum of 4,000 characters. If your prompt is longer than this limit, we suggest using SSML for Text-to-Speech-based play actions.
61
+
- Speech input for recordings is captured for 1:1 calls but not recorded in group calls when recording is enabled.
62
+
- Speech service quota increases can be requested if you exceed your quota limit. Follow the steps outlined [here](/azure/ai-services/speech-service/speech-services-quotas-and-limits) to request an increase.
0 commit comments