Skip to content

Commit cf05d16

Browse files
authored
Merge pull request #5790 from eric-urban/eur/voice-live-api
voice live api regions and pricing and foundry
2 parents 70a36ed + 653100c commit cf05d16

File tree

14 files changed

+317
-143
lines changed

14 files changed

+317
-143
lines changed

articles/ai-services/speech-service/includes/how-to/video-translation/ai-foundry.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -68,6 +68,8 @@ To create a video translation project, follow these steps:
6868

6969
:::image type="content" source="../../../media/video-translation/foundry-portal/capabilities-by-scenario.png" alt-text="Screenshot of filtering Speech service capabilities by scenario." lightbox="../../../media/video-translation/foundry-portal/capabilities-by-scenario.png":::
7070

71+
1. Select **Go to Speech playground**.
72+
7173
1. Select **My videos** > **Upload video**.
7274

7375
:::image type="content" source="../../../media/video-translation/foundry-portal/select-upload-video.png" alt-text="Screenshot of selecting the upload video button." lightbox="../../../media/video-translation/foundry-portal/select-upload-video.png":::
Lines changed: 51 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,51 @@
1+
---
2+
title: include file
3+
description: include file
4+
author: eric-urban
5+
ms.author: eur
6+
ms.service: azure-ai-speech
7+
ms.topic: include
8+
ms.date: 7/1/2025
9+
ms.custom: references_regions
10+
---
11+
12+
In this article, you learn how to use voice live with generative AI and Azure AI Speech in the [Azure AI Foundry portal](https://ai.azure.com/?cid=learnDocs).
13+
14+
## Prerequisites
15+
16+
- An Azure subscription. <a href="https://azure.microsoft.com/free/ai-services" target="_blank">Create one for free</a>.
17+
- <a href="https://www.python.org/" target="_blank">Python 3.8 or later version</a>. We recommend using Python 3.10 or later, but having at least Python 3.8 is required. If you don't have a suitable version of Python installed, you can follow the instructions in the [VS Code Python Tutorial](https://code.visualstudio.com/docs/python/python-tutorial#_install-a-python-interpreter) for the easiest way of installing Python on your operating system.
18+
- An [Azure AI Foundry resource](../../../../multi-service-resource.md) created in one of the supported regions. For more information about region availability, see the [Voice Live API overview documentation](../../../voice-live.md).
19+
20+
> [!TIP]
21+
> To use the Voice Live API, you don't need to deploy an audio model with your Azure AI Foundry resource. The Voice Live API is fully managed, and the model is automatically deployed for you. For more information about models availability, see the [Voice Live API overview documentation](../../../voice-live.md).
22+
23+
## Try out voice live in the Speech playground
24+
25+
To try out the voice live demo, follow these steps:
26+
27+
1. Go to your project in [Azure AI Foundry](https://ai.azure.com/?cid=learnDocs).
28+
1. Select **Playgrounds** from the left pane.
29+
1. In the **Speech playground** tile, select **Try the Speech playground**.
30+
1. Select **Speech capabilities by scenario** > **Voice live**.
31+
32+
:::image type="content" source="../../../media/voice-live/foundry-portal/capabilities-by-scenario.png" alt-text="Screenshot of filtering Speech service capabilities by scenario." lightbox="../../../media/voice-live/foundry-portal/capabilities-by-scenario.png":::
33+
34+
1. Select a sample scenario, such as **Casual chat**.
35+
36+
:::image type="content" source="../../../media/voice-live/foundry-portal/casual-chat-start.png" alt-text="Screenshot of selecting the casual chat example scenario in the Speech playground." lightbox="../../../media/voice-live/foundry-portal/casual-chat-start.png":::
37+
38+
1. Select **Start** to start chatting with the chat agent.
39+
40+
1. Select **End** to end the chat session.
41+
42+
1. Select a new generative AI model from the drop-down list via **Configuration** > **GenAI** > **Generative AI model**.
43+
44+
> [!NOTE]
45+
> You can also select an agent that you configured in the **Agents** playground.
46+
47+
:::image type="content" source="../../../media/voice-live/foundry-portal/casual-chat-generative-ai-select.png" alt-text="Screenshot of the casual chat example scenario in the Speech playground." lightbox="../../../media/voice-live/foundry-portal/casual-chat-generative-ai-select.png":::
48+
49+
1. Edit other settings as needed, such as the **Response instructions**, **Voice**, and **Speaking rate**.
50+
51+
1. Select **Start** to start speaking again and select **End** to end the chat session.

articles/ai-services/speech-service/includes/quickstarts/voice-live-api/realtime-python.md renamed to articles/ai-services/speech-service/includes/quickstarts/voice-live-api/python.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ author: eric-urban
44
ms.author: eur
55
ms.service: azure-ai-openai
66
ms.topic: include
7-
ms.date: 6/27/2025
7+
ms.date: 7/1/2025
88
---
99

1010
## Prerequisites

articles/ai-services/speech-service/includes/quickstarts/voice-live-api/resource-authentication.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,9 @@
11
---
22
author: eric-urban
33
ms.author: eur
4-
ms.service: azure-ai-openai
4+
ms.service: azure-ai-speech
55
ms.topic: include
6-
ms.date: 5/19/2025
6+
ms.date: 7/1/2025
77
---
88

99
You need to retrieve the following information to authenticate your application with your Azure AI Foundry resource:
-55 KB
Loading
-61.7 KB
Loading
135 KB
Loading
273 KB
Loading
279 KB
Loading

articles/ai-services/speech-service/regions.md

Lines changed: 47 additions & 34 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ author: eric-urban
66
manager: nitinme
77
ms.service: azure-ai-speech
88
ms.topic: conceptual
9-
ms.date: 3/10/2025
9+
ms.date: 7/1/2025
1010
ms.author: eur
1111
ms.custom: references_regions
1212
#Customer intent: As a developer, I want to learn about the available regions and endpoints for the Speech service.
@@ -172,6 +172,20 @@ The regions in these tables support most of the core features of the Speech serv
172172
| westus2 ||
173173
| westus3 ||
174174

175+
# [Voice live](#tab/voice-live)
176+
177+
| **Region** | **gpt-4o-realtime-preview** | **gpt-4o-mini-realtime-preview** | **gpt-4o** | **gpt-4o-mini** | **gpt-4.1** | **gpt-4.1-mini** | **gpt-4.1-nano** | **phi4-mm-realtime** | **phi4-mini** |
178+
|-----|-----|-----|-----|-----|-----|-----|-----|-----|-----|
179+
| centralindia | Cross-region<sup>1</sup> | Cross-region<sup>1</sup> | Global standard | Global standard | - | Global standard | Global standard | - | - |
180+
| eastus2 | Global standard | Global standard | Data zone standard | Data zone standard | Data zone standard | Data zone standard | Data zone standard | Regional | Regional |
181+
| southeastasia | - | - | - | - | - | Global standard | Global standard | - | - |
182+
| swedencentral | Global standard | Global standard | Data zone standard | Data zone standard | Data zone standard | Data zone standard | Data zone standard | Regional | Regional |
183+
| westus2 | Cross-region<sup>2</sup> | Cross-region<sup>2</sup> | Data zone standard | Data zone standard | Data zone standard | Data zone standard | Data zone standard | Regional | Regional |
184+
185+
<sup>1</sup> The Azure AI Foundry resource must be in Central India. Azure AI Speech features remain in Central India. The voice live API uses Sweden Central as needed for generative AI load balancing.
186+
187+
<sup>2</sup> The Azure AI Foundry resource must be in West US 2. Azure AI Speech features remain in West US 2. The voice live API uses East US 2 as needed for generative AI load balancing.
188+
175189
# [Intent recognition](#tab/intent-recognition)
176190

177191
| **Region** | **Intent recognition** |
@@ -246,41 +260,40 @@ The regions in these tables support most of the core features of the Speech serv
246260
| westus2 |||
247261
| westus3 | | |
248262

249-
250263
# [Scenarios](#tab/scenarios)
251264

252-
| **Region** | **Pronunciation assessment** | **Speaker recognition** | **Voice assistants** | **Voice Live API** |
253-
|-----|-----|-----|-----|
254-
| australiaeast ||| | |
255-
| brazilsouth || | | |
256-
| canadacentral |||| |
257-
| centralindia ||| | |
258-
| centralus ||| | |
259-
| eastasia |||| |
260-
| eastus |||| |
261-
| eastus2 |||||
262-
| francecentral ||| | |
263-
| germanywestcentral ||| | |
264-
| japaneast ||| | |
265-
| japanwest ||| | |
266-
| koreacentral ||| | |
267-
| northcentralus || | | |
268-
| northeurope |||| |
269-
| norwayeast ||| | |
270-
| qatarcentral ||| | |
271-
| southafricanorth || | | |
272-
| southcentralus || || |
273-
| southeastasia |||| |
274-
| swedencentral ||| ||
275-
| switzerlandnorth || | | |
276-
| switzerlandwest ||| | |
277-
| uaenorth || | | |
278-
| uksouth ||| | |
279-
| westcentralus |||| |
280-
| westeurope |||| |
281-
| westus |||| |
282-
| westus2 |||| |
283-
| westus3 ||| | |
265+
| **Region** | **Pronunciation assessment** | **Speaker recognition** | **Voice assistants** |
266+
|-----|-----|-----|
267+
| australiaeast ||| |
268+
| brazilsouth || | |
269+
| canadacentral ||||
270+
| centralindia ||| |
271+
| centralus ||| |
272+
| eastasia ||||
273+
| eastus ||||
274+
| eastus2 ||||
275+
| francecentral ||| |
276+
| germanywestcentral ||| |
277+
| japaneast ||| |
278+
| japanwest ||| |
279+
| koreacentral ||| |
280+
| northcentralus || | |
281+
| northeurope ||||
282+
| norwayeast ||| |
283+
| qatarcentral ||| |
284+
| southafricanorth || | |
285+
| southcentralus || ||
286+
| southeastasia ||||
287+
| swedencentral ||| |
288+
| switzerlandnorth || | |
289+
| switzerlandwest ||| |
290+
| uaenorth || | |
291+
| uksouth ||| |
292+
| westcentralus ||||
293+
| westeurope ||||
294+
| westus ||||
295+
| westus2 ||||
296+
| westus3 ||| |
284297

285298
> [!NOTE]
286299
> The [Speech SDK](speech-sdk.md) supports voice assistant capabilities through Direct Line Speech.

0 commit comments

Comments
 (0)