Skip to content

Commit 439433e

Browse files
Merge pull request #300799 from valindrae/valindrae-ga
Valindrae ga
2 parents 3c50958 + 076d9f3 commit 439433e

File tree

5 files changed

+32
-41
lines changed

5 files changed

+32
-41
lines changed

articles/communication-services/concepts/call-automation/azure-communication-services-azure-cognitive-services-integration.md

Lines changed: 16 additions & 29 deletions
Original file line numberDiff line numberDiff line change
@@ -14,27 +14,32 @@ services: azure-communication-services
1414

1515
# Connect Azure Communication Services with Azure AI services
1616

17-
Azure Communication Services Call Automation APIs provide developers the ability to steer and control the Azure Communication Services Telephony, VoIP or WebRTC calls using real-time event triggers to perform actions based on custom business logic specific to their domain. Within the Call Automation APIs developers can use simple AI powered APIs, which can be used to play personalized greeting messages, recognize conversational voice inputs to gather information on contextual questions to drive a more self-service model with customers, use sentiment analysis to improve customer service overall. These content specific APIs are orchestrated through **Azure AI Services** with support for customization of AI models without developers needing to terminate media streams on their services and streaming back to Azure for AI functionality.
17+
Azure Communication Services Call Automation APIs allow developers to steer and control calls made through Azure Communication Services, including telephony, VoIP, and WebRTC. These APIs use real-time event triggers, which enable actions based on custom business logic that is specific to each developer’s domain. With Call Automation APIs, developers can use simple AI-powered features. For example, they can play personalized greetings, recognize spoken responses to gather information from customers, and analyze sentiment to improve service. These targeted APIs are managed through **Azure AI Foundry**, which allow developers to customize AI models. Importantly, developers do not need to deal with media streams or send them back to Azure for these AI functions—the processing happens seamlessly.
1818

19-
All this is possible with one-click where enterprises can access a secure solution and link their models through the portal. Furthermore, developers and enterprises don't need to manage credentials. Connecting your Azure AI services uses managed identities to access user-owned resources. Developers can use managed identities to authenticate any resource that supports Microsoft Entra authentication.
19+
All this functionality is possible with one click, allowing enterprises to access a secure solution and link their models through the portal. Furthermore, developers and enterprises don't need to manage credentials. Connecting your Azure AI services uses managed identities to access user-owned resources. Developers can use managed identities to authenticate any resource that supports Microsoft Entra authentication.
2020

2121
Azure AI services can be easily integrated into any application regardless of the programming language. When creating an Azure Resource in Azure portal, enable the option and provide the URL to the Azure AI services. This simple experience allows developers to meet their needs, scale, and avoid investing time and resources into designing and maintaining a custom solution.
2222

2323
> [!NOTE]
24-
> This integration is supported in limited regions for Azure AI services, for more information about which regions are supported please view the limitations section at the bottom of this document. This integration only supports Multi-service Cognitive Service resource, we recommend if you're creating a new Azure AI Service resource you create a Multi-service Cognitive Service resource or when you're connecting an existing resource confirm that it is a Multi-service Cognitive Service resource.
24+
> This integration only supports Multi-service Cognitive Service resource, we recommend if you're creating a new Azure AI Service resource you create a Multi-service Cognitive Service resource or when you're connecting an existing resource confirm that it's a Multi-service Cognitive Service resource.
2525
2626
## Common use cases
2727

2828
### Build applications that can play and recognize speech
2929

30-
With the ability to connect your Azure AI services to Azure Communication Services. You can enable custom play functionality, using [Text-to-Speech](../../../../articles/cognitive-services/Speech-Service/text-to-speech.md) and [Speech Synthesis Markup Language (SSML)](../../../../articles/cognitive-services/Speech-Service/speech-synthesis-markup.md) configuration, to play more customized and natural sounding audio to users. Through the Azure AI services connection, you can also use the Speech-To-Text service to incorporate recognition of voice responses that can be converted into actionable tasks through business logic in the application. These functions can be further enhanced through the ability to create custom models within Azure AI services that are bespoke to your domain and region, through the ability to choose which languages are spoken and recognized, custom voices and custom models built based on your experience.
30+
With the ability to connect your Azure AI services to Azure Communication Services. You can enable custom play functionality, using [Text-to-Speech](../../../../articles/cognitive-services/Speech-Service/text-to-speech.md) and [Speech Synthesis Markup Language (SSML)](../../../../articles/cognitive-services/Speech-Service/speech-synthesis-markup.md) configuration, to play more customized and natural sounding audio to users. Through the Azure AI services connection, you can also use the Speech-To-Text service to incorporate recognition of voice responses that can be converted into actionable tasks through business logic in the application. These functions can be further enhanced within Azure AI services by:
31+
32+
- Creating custom models tailored to your domain and region
33+
- Selecting which languages are spoken and recognized
34+
- Designing custom voices
35+
- Building additional models based on your experience
3136

3237
## Runtime flow
3338
[![Screen shot of integration runtime flow.](./media/run-time-flow.png)](./media/run-time-flow.png#lightbox)
3439

3540
## Azure portal experience
36-
You'll need to connect your Azure Communication Services resource with the Azure AI resource through the Azure portal. There are two ways you can accomplish this step:
37-
- By navigating through the steps of the Cognitive Services tab in your Azure Communication Services (recommended).
41+
You need to connect your Azure Communication Services resource with the Azure AI resource through the Azure portal. There are two ways you can accomplish this step:
42+
- Navigating through the steps of the Cognitive Services tab in your Azure Communication Services (recommended).
3843
- Manually adding the Managed Identity to your Azure Communication Services resource. This step is more advanced and requires a little more effort to connect your Azure Communication Services to your Azure AI services.
3944

4045
## Prerequisites
@@ -46,7 +51,7 @@ You'll need to connect your Azure Communication Services resource with the Azure
4651
### Connecting through the Azure portal
4752

4853
1. Open your Azure Communication Services resource and click on the Cognitive Services tab.
49-
2. If system-assigned managed identity isn't enabled, you'll need to enable it.
54+
2. If system-assigned managed identity isn't enabled, you need to enable it.
5055
3. In the Cognitive Services tab, click on "Enable Managed Identity" button.
5156

5257
[![Screenshot of Enable Managed Identity button.](./media/enabled-identity.png)](./media/enabled-identity.png#lightbox)
@@ -57,7 +62,7 @@ You'll need to connect your Azure Communication Services resource with the Azure
5762
5. Once the identity is enabled, you should see something similar.
5863
[![Screenshot of enabled identity.](./media/identity-saved.png)](./media/identity-saved.png#lightbox)
5964

60-
6. When managed identity is enabled the Cognitive Service tab should show a button 'Connect cognitive service' to connect the two services.
65+
6. When managed identity is enabled, the Cognitive Service tab should show a button 'Connect cognitive service' to connect the two services.
6166
[![Screenshot of Connect cognitive services button.](./media/cognitive-services.png)](./media/cog-svc.png#lightbox)
6267

6368
7. Click on 'Connect cognitive service', select the Subscription, Resource Group and Resource and click 'Connect' in the context pane that opens up.
@@ -109,27 +114,9 @@ Your Azure Communication Service has now been linked to your Azure Cognitive Ser
109114

110115
## Azure AI services regions supported
111116

112-
This integration between Azure Communication Services and Azure AI services is only supported in the following regions:
113-
- centralus
114-
- northcentralus
115-
- southcentralus
116-
- westcentralus
117-
- eastus
118-
- eastus2
119-
- westus
120-
- westus2
121-
- westus3
122-
- canadacentral
123-
- northeurope
124-
- westeurope
125-
- uksouth
126-
- southafricanorth
127-
- centralindia
128-
- eastasia
129-
- southeastasia
130-
- australiaeast
131-
- brazilsouth
132-
- uaenorth
117+
Our integration between Azure Communication Services (ACS) and Azure AI is fully aligned with the regional availability of Azure AI Foundry. This means that ACS to Azure AI integration is supported in all regions where Azure AI Foundry is available.
118+
119+
For the most up-to-date list of supported regions, refer to the official Azure AI Foundry [documentation](/azure/ai-services/speech-service/regions).
133120

134121
## Known limitations
135122

articles/communication-services/concepts/call-automation/play-action.md

Lines changed: 7 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -12,15 +12,15 @@ ms.author: kpunjabi
1212
# Playing audio in call
1313

1414
The play action provided through the Azure Communication Services Call Automation SDK allows you to play audio prompts to participants in the call. This action can be accessed through the server-side implementation of your application. You can play audio to call participants through one of two methods;
15-
- Providing Azure Communication Services access to prerecorded audio files of WAV format, that Azure Communication Services can access with support for authentication
15+
- Providing Azure Communication Services access to prerecorded audio files in WAV format, which Azure Communication Services can access with support for authentication.
1616
- Regular text that can be converted into speech output through the integration with Azure AI services.
1717

1818
You can use the newly announced integration between [Azure Communication Services and Azure AI services](./azure-communication-services-azure-cognitive-services-integration.md) to play personalized responses using Azure [Text-To-Speech](../../../../articles/cognitive-services/Speech-Service/text-to-speech.md). You can use human like prebuilt neural voices out of the box or create custom neural voices that are unique to your product or brand. For more information on supported voices, languages and locales see [Language and voice support for the Speech service](../../../../articles/cognitive-services/Speech-Service/language-support.md).
1919
> [!NOTE]
20-
> Azure Communication Services currently supports two file formats, MP3 files with ID3V2TAG and WAV files formatted as 16-bit PCM mono channel audio recorded at 16KHz. You can create your own audio files using [Speech synthesis with Audio Content Creation tool](/azure/ai-services/speech-service/how-to-audio-content-creation).
20+
> Azure Communication Services currently supports two file formats, MP3 files with ID3V2TAG and WAV files formatted as 16-bit Pulse Code Modulation (PCM) mono channel audio recorded at 16 KHz. You can create your own audio files using [Speech synthesis with Audio Content Creation tool](/azure/ai-services/speech-service/how-to-audio-content-creation).
2121
2222
## Prebuilt Neural Text to Speech voices
23-
Microsoft uses deep neural networks to overcome the limits of traditional speech synthesis with regard to stress and intonation in spoken language. Prosody prediction and voice synthesis occur simultaneously, resulting in a more fluid and natural sounding output. You can use these neural voices to make interactions with your chatbots and voice assistants more natural and engaging. There are over 100 prebuilt voices to choose from. Learn more about [Azure Text-to-Speech voices](../../../../articles/cognitive-services/Speech-Service/language-support.md).
23+
Microsoft uses advanced machine learning to overcome the limitations of traditional speech synthesis in handling stress and intonation in spoken language. Prosody prediction and voice synthesis occur simultaneously, resulting in a more fluid and natural sounding output. You can use these neural voices to make interactions with your chatbots and voice assistants more natural and engaging. There are over 100 prebuilt voices to choose from. Learn more about [Azure Text-to-Speech voices](../../../../articles/cognitive-services/Speech-Service/language-support.md).
2424

2525
## Common use cases
2626

@@ -31,13 +31,13 @@ Your application might want to play some sort of announcement when a participant
3131

3232
### Self-serve customers
3333

34-
In scenarios with IVRs and virtual assistants, you can use your application or bots to play audio prompts to callers, this prompt can be in the form of a menu to guide the caller through their interaction.
34+
In scenarios with Interactive Voice Response (IVR) systems and virtual assistants, you can use your application or bots to play audio prompts to callers, this prompt can be in the form of a menu to guide the caller through their interaction.
3535

3636
### Hold music
3737
The play action can also be used to play hold music for callers. This action can be set up in a loop so that the music keeps playing until an agent is available to assist the caller.
3838

3939
### Playing compliance messages
40-
As part of compliance requirements in various industries, vendors are expected to play legal or compliance messages to callers, for example, This call is recorded for quality purposes.”.
40+
As part of compliance requirements in various industries, vendors are expected to play legal or compliance messages to callers, for example, "This call is recorded for quality purposes."
4141

4242
## Sample architecture for playing audio in call using Text-To-Speech
4343

@@ -48,7 +48,8 @@ As part of compliance requirements in various industries, vendors are expected t
4848
![Screenshot of flow for play action.](./media/play-action.png)
4949

5050
## Known limitations
51-
- Text-to-Speech text prompts support a maximum of 400 characters, if your prompt is longer than this we suggest using SSML for Text-to-Speech based play actions.
51+
- Text-to-Speech prompts support a maximum of 4,000 characters. If your prompt is longer than this limit, we suggest using Speech Synthesis Markup Language (SSML) for Text-to-Speech based play actions.
52+
- HD Voices from Azure AI Foundry aren't officially supported. Using these voices results in timeouts.
5253
- For scenarios where you exceed your Speech service quota limit, you can request to increase this limit by following the steps outlined [here](/azure/ai-services/speech-service/speech-services-quotas-and-limits).
5354

5455
## Next Steps

articles/communication-services/concepts/call-automation/recognize-action.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -66,7 +66,8 @@ The recognize action can be used for many reasons, here are a few examples of ho
6666

6767
## Known limitation
6868
- In-band DTMF is not supported, use RFC 2833 DTMF instead.
69-
- Text-to-Speech text prompts support a maximum of 400 characters, if your prompt is longer than this we suggest using SSML for Text-to-Speech based play actions.
69+
- Text-to-Speech text prompts support a maximum of 4,000 characters, if your prompt is longer than this we suggest using SSML for Text-to-Speech based play actions.
70+
- When recording is enabled speech input is recorded for 1:1 calls, but isn't recorded in group calls.
7071
- For scenarios where you exceed your Speech service quota limit, you can request to increase this limit by following the steps outlined [here](/azure/ai-services/speech-service/speech-services-quotas-and-limits).
7172

7273
## Next steps

articles/communication-services/how-tos/call-automation/play-action.md

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ zone_pivot_groups: acs-js-csharp-java-python
1414

1515
# Customize voice prompts to users with Play action
1616

17-
This guide will help you get started with playing audio files to participants by using the play action provided through Azure Communication Services Call Automation SDK.
17+
This guide helps you get started with playing audio files to participants by using the play action provided through Azure Communication Services Call Automation SDK.
1818

1919
::: zone pivot="programming-language-csharp"
2020
[!INCLUDE [Play audio with .NET](./includes/play-audio-quickstart-csharp.md)]
@@ -48,8 +48,9 @@ This guide will help you get started with playing audio files to participants by
4848
|PlayFailed | 500 | 8572 | Action failed due to play service shutdown. |
4949

5050
## Known limitations
51-
- Text-to-Speech text prompts support a maximum of 400 characters, if your prompt is longer than this we suggest using SSML for Text-to-Speech based play actions.
52-
- For scenarios where you exceed your Speech service quota limit, you can request to increase this limit by following the steps outlined [here](/azure/ai-services/speech-service/speech-services-quotas-and-limits).
51+
- Text-to-Speech prompts support a maximum of 4,000 characters. If your prompt is longer than this limit, we suggest using SSML for Text-to-Speech-based play actions.
52+
- HD Voices from Azure AI Foundry aren't officially supported. These will result in timeouts if used.
53+
- Speech service quota increases can be requested if you exceed your quota limit. Follow the steps outlined [here](/azure/ai-services/speech-service/speech-services-quotas-and-limits) to request an increase.
5354

5455
## Clean up resources
5556

articles/communication-services/how-tos/call-automation/recognize-action.md

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -57,8 +57,9 @@ This guide helps you get started recognizing DTMF input provided by participants
5757

5858
## Known limitations
5959
- In-band DTMF isn't supported. Use RFC 2833 DTMF instead.
60-
- Text-to-Speech text prompts support a maximum of 400 characters, if your prompt is longer than this we suggest using SSML for Text-to-Speech based play actions.
61-
- For scenarios where you exceed your Speech service quota limit, you can request to increase this limit by following the steps outlined in [Speech services quotas and limits](/azure/ai-services/speech-service/speech-services-quotas-and-limits).
60+
- Text-to-Speech prompts support a maximum of 4,000 characters. If your prompt is longer than this limit, we suggest using SSML for Text-to-Speech-based play actions.
61+
- Speech input for recordings is captured for 1:1 calls but not recorded in group calls when recording is enabled.
62+
- Speech service quota increases can be requested if you exceed your quota limit. Follow the steps outlined [here](/azure/ai-services/speech-service/speech-services-quotas-and-limits) to request an increase.
6263

6364
## Clean up resources
6465

0 commit comments

Comments
 (0)