Skip to content

Commit 3ca067b

Browse files
committed
Update procuct naming for consistency.
1 parent ac0fa7b commit 3ca067b

File tree

9 files changed

+69
-69
lines changed

9 files changed

+69
-69
lines changed

articles/ai-services/speech-service/includes/quickstarts/voice-live-agents/intro.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16,4 +16,4 @@ You can create and run an application to use voice live with agents for real-tim
1616

1717
- This separation also supports better maintainability and scalability for scenarios where multiple conversational experiences or business logic variations are needed.
1818

19-
To instead use the voice live API without agents, see the [voice live quickstart](/azure/ai-services/speech-service/voice-live-quickstart).
19+
To instead use the Voice live API without agents, see the [Voice live API quickstart](/azure/ai-services/speech-service/voice-live-quickstart).

articles/ai-services/speech-service/includes/quickstarts/voice-live-api/intro.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,9 @@
11
---
2-
author: PatrickFarley
2+
author: goergenj
33
ms.service: azure-ai-speech
44
ms.topic: include
5-
ms.date: 7/31/2025
6-
ms.author: pafarley
5+
ms.date: 9/26/2025
6+
ms.author: jagoerge
77
---
88

99
You create and run an application to use voice live directly with generative AI models for real-time voice agents.
@@ -16,4 +16,4 @@ You create and run an application to use voice live directly with generative AI
1616

1717
- Direct model use is suitable for scenarios where agent-level abstraction or built-in logic is unnecessary.
1818

19-
To instead use the voice live API with agents, see the [voice live agents quickstart](/azure/ai-services/speech-service/voice-live-agents-quickstart).
19+
To instead use the Voice live API with agents, see the [Voice live API agents quickstart](/azure/ai-services/speech-service/voice-live-agents-quickstart).

articles/ai-services/speech-service/regions.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ author: goergenj
66
manager: nitinme
77
ms.service: azure-ai-speech
88
ms.topic: conceptual
9-
ms.date: 9/16/2025
9+
ms.date: 9/26/2025
1010
ms.author: jagoerge
1111
ms.custom: references_regions
1212
#Customer intent: As a developer, I want to learn about the available regions and endpoints for the Speech service.
@@ -187,9 +187,9 @@ The regions in these tables support most of the core features of the Speech serv
187187
|uksouth| - | - | Global standard | Global standard | Global standard | Global standard | - | - | - | - | - | - |
188188
|westeurope| - | - | Data zone standard | Data zone standard | Data zone standard | Data zone standard | - | - | - | - | - | - |
189189

190-
<sup>1</sup> The Azure AI Foundry resource must be in Central India. Azure AI Speech features remain in Central India. The voice live API uses Sweden Central as needed for generative AI load balancing.
190+
<sup>1</sup> The Azure AI Foundry resource must be in Central India. Azure AI Speech features remain in Central India. The Voice live API uses Sweden Central as needed for generative AI load balancing.
191191

192-
<sup>2</sup> The Azure AI Foundry resource must be in West US 2. Azure AI Speech features remain in West US 2. The voice live API uses East US 2 as needed for generative AI load balancing.
192+
<sup>2</sup> The Azure AI Foundry resource must be in West US 2. Azure AI Speech features remain in West US 2. The Voice live API uses East US 2 as needed for generative AI load balancing.
193193

194194
# [Intent recognition](#tab/intent-recognition)
195195

articles/ai-services/speech-service/voice-live-agents-quickstart.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,5 +28,5 @@ recommendations: false
2828
## Related content
2929

3030
- Try the [Voice live quickstart](./voice-live-quickstart.md)
31-
- Learn more about [How to use the voice live API](./voice-live-how-to.md)
31+
- Learn more about [How to use the Voice live API](./voice-live-how-to.md)
3232
- See the [Voice live API reference](./voicelive-api-reference.md)

articles/ai-services/speech-service/voice-live-how-to-customize.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,15 +1,15 @@
11
---
22
title: How to customize voice live input and output
33
titleSuffix: Azure AI services
4-
description: Learn how to use the voice live API with customized models.
4+
description: Learn how to use the Voice live API with customized models.
55
manager: nitinme
66
author: goergenj
77
ms.author: jagoerge
88
ms.service: azure-ai-speech
99
ms.topic: how-to
10-
ms.date: 9/25/2025
10+
ms.date: 9/26/2025
1111
ms.custom: custom speech, custom voice, custom avatar, fine-tuning
12-
# Customer intent: As a developer, I want to learn how to use custom models with the voice live API for real-time voice agents.
12+
# Customer intent: As a developer, I want to learn how to use custom models with the Voice live API for real-time voice agents.
1313
---
1414

1515
# How to customize voice live input and output
@@ -69,7 +69,7 @@ Example session configuration with custom speech models. In this case, if the de
6969
```
7070

7171
> [!NOTE]
72-
> In order to use a custom speech model with voice live API, the model must be available on the same Azure AI Foundry resource you are using to call the voice live API. If you trained the model on a different Azure AI Foundry or Azure AI Speech resource you have to copy the model to the resource you are using to call the voice live API.
72+
> In order to use a custom speech model with Voice live API, the model must be available on the same Azure AI Foundry resource you are using to call the Voice live API. If you trained the model on a different Azure AI Foundry or Azure AI Speech resource you have to copy the model to the resource you are using to call the Voice live API.
7373
> You pay separately for custom speech training and model hosting.
7474
7575
## Speech output customization
@@ -105,21 +105,21 @@ You can use a custom voice for audio output. For information about how to create
105105
```
106106

107107
> [!NOTE]
108-
> In order to use a custom voice model with voice live API, the model must be available on the same Azure AI Foundry resource you are using to call the voice live API. If you trained the model on a different Azure AI Foundry or Azure AI Speech resource you have to copy the model to the resource you are using to call the voice live API.
108+
> In order to use a custom voice model with Voice live API, the model must be available on the same Azure AI Foundry resource you are using to call the Voice live API. If you trained the model on a different Azure AI Foundry or Azure AI Speech resource you have to copy the model to the resource you are using to call the Voice live API.
109109
> You pay separately for custom voice training and model hosting.
110110
111111
### Azure custom avatar
112112

113113
[Text to speech avatar](./text-to-speech-avatar/what-is-text-to-speech-avatar.md) converts text into a digital video of a photorealistic human (either a standard avatar or a [custom text to speech avatar](./text-to-speech-avatar/what-is-custom-text-to-speech-avatar.md)) speaking with a natural-sounding voice.
114114

115-
The configuration for a custom avatar does not differ from the configuration of a standard avatar. Please refer to [How to use the voice live API - Azure text to speech avatar](./voice-live-how-to.md#azure-text-to-speech-avatar) for a detailed example.
115+
The configuration for a custom avatar does not differ from the configuration of a standard avatar. Please refer to [How to use the Voice live API - Azure text to speech avatar](./voice-live-how-to.md#azure-text-to-speech-avatar) for a detailed example.
116116

117117
> [!NOTE]
118-
> In order to use a custom voice model with voice live API, the model must be available on the same Azure AI Foundry resource you are using to call the voice live API. If you trained the model on a different Azure AI Foundry or Azure AI Speech resource you have to copy the model to the resource you are using to call the voice live API.
118+
> In order to use a custom voice model with Voice live API, the model must be available on the same Azure AI Foundry resource you are using to call the Voice live API. If you trained the model on a different Azure AI Foundry or Azure AI Speech resource you have to copy the model to the resource you are using to call the Voice live API.
119119
> You pay separately for custom avatar training and model hosting.
120120
121121

122122
## Related content
123123

124124
- Try out the [Voice live API quickstart](./voice-live-quickstart.md)
125-
- Learn more about [How to use the voice live API](./voice-live-how-to.md)
125+
- Learn more about [How to use the Voice live API](./voice-live-how-to.md)

articles/ai-services/speech-service/voice-live-how-to.md

Lines changed: 16 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -1,43 +1,43 @@
11
---
2-
title: How to use the voice live API (Preview)
2+
title: How to use the Voice live API (Preview)
33
titleSuffix: Azure AI services
4-
description: Learn how to use the voice live API for real-time voice agents.
4+
description: Learn how to use the Voice live API for real-time voice agents.
55
manager: nitinme
66
author: goergenj
77
ms.author: jagoerge
88
ms.service: azure-ai-speech
99
ms.topic: how-to
10-
ms.date: 9/16/2025
10+
ms.date: 9/26/2025
1111
ms.custom: references_regions
12-
# Customer intent: As a developer, I want to learn how to use the voice live API for real-time voice agents.
12+
# Customer intent: As a developer, I want to learn how to use the Voice live API for real-time voice agents.
1313
---
1414

15-
# How to use the voice live API (Preview)
15+
# How to use the Voice live API (Preview)
1616

1717
[!INCLUDE [Feature preview](./includes/previews/preview-generic.md)]
1818

19-
The voice live API provides a capable WebSocket interface compared to the [Azure OpenAI Realtime API](../../ai-foundry/openai/how-to/realtime-audio.md).
19+
The Voice live API provides a capable WebSocket interface compared to the [Azure OpenAI Realtime API](../../ai-foundry/openai/how-to/realtime-audio.md).
2020

21-
Unless otherwise noted, the voice live API uses the [same events](/azure/ai-foundry/openai/realtime-audio-reference?context=/azure/ai-services/speech-service/context/context) as the Azure OpenAI Realtime API. This document provides a reference for the event message properties that are specific to the voice live API.
21+
Unless otherwise noted, the Voice live API uses the [same events](/azure/ai-foundry/openai/realtime-audio-reference?context=/azure/ai-services/speech-service/context/context) as the Azure OpenAI Realtime API. This document provides a reference for the event message properties that are specific to the Voice live API.
2222

2323
## Supported models and regions
2424

25-
For a table of supported models and regions, see the [voice live API overview](./voice-live.md#supported-models-and-regions).
25+
For a table of supported models and regions, see the [Voice live API overview](./voice-live.md#supported-models-and-regions).
2626

2727
## Authentication
2828

29-
An [Azure AI Foundry resource](../multi-service-resource.md) is required to access the voice live API.
29+
An [Azure AI Foundry resource](../multi-service-resource.md) is required to access the Voice live API.
3030

3131
### WebSocket endpoint
3232

33-
The WebSocket endpoint for the voice live API is `wss://<your-ai-foundry-resource-name>.services.ai.azure.com/voice-live/realtime?api-version=2025-10-01` or, for older resources, `wss://<your-ai-foundry-resource-name>.cognitiveservices.azure.com/voice-live/realtime?api-version=2025-10-01`.
33+
The WebSocket endpoint for the Voice live API is `wss://<your-ai-foundry-resource-name>.services.ai.azure.com/voice-live/realtime?api-version=2025-10-01` or, for older resources, `wss://<your-ai-foundry-resource-name>.cognitiveservices.azure.com/voice-live/realtime?api-version=2025-10-01`.
3434
The endpoint is the same for all models. The only difference is the required `model` query parameter, or, when using the Agent service, the `agent_id` and `project_id` parameters.
3535

3636
For example, an endpoint for a resource with a custom domain would be `wss://<your-ai-foundry-resource-name>.services.ai.azure.com/voice-live/realtime?api-version=2025-10-01&model=gpt-realtime`
3737

3838
### Credentials
3939

40-
The voice live API supports two authentication methods:
40+
The Voice live API supports two authentication methods:
4141

4242
- **Microsoft Entra** (recommended): Use token-based authentication for an Azure AI Foundry resource. Apply a retrieved authentication token using a `Bearer` token with the `Authorization` header.
4343
- **API key**: An `api-key` can be provided in one of two ways:
@@ -52,7 +52,7 @@ For the recommended keyless authentication with Microsoft Entra ID, you need to:
5252

5353
## Session configuration
5454

55-
Often, the first event sent by the caller on a newly established voice live API session is the [`session.update`](../openai/realtime-audio-reference.md?context=/azure/ai-services/speech-service/context/context#realtimeclienteventsessionupdate) event. This event controls a wide set of input and output behavior, with output and response generation properties then later overridable using the [`response.create`](../openai/realtime-audio-reference.md?context=/azure/ai-services/speech-service/context/context#realtimeclienteventresponsecreate) event.
55+
Often, the first event sent by the caller on a newly established Voice live API session is the [`session.update`](../openai/realtime-audio-reference.md?context=/azure/ai-services/speech-service/context/context#realtimeclienteventsessionupdate) event. This event controls a wide set of input and output behavior, with output and response generation properties then later overridable using the [`response.create`](../openai/realtime-audio-reference.md?context=/azure/ai-services/speech-service/context/context#realtimeclienteventresponsecreate) event.
5656

5757
Here's an example `session.update` message that configures several aspects of the session, including turn detection, input audio processing, and voice output. Most session parameters are optional and can be omitted if not needed.
5858

@@ -91,7 +91,7 @@ The server responds with a [`session.updated`](../openai/realtime-audio-referenc
9191
The following sections describe the properties of the `session` object that can be configured in the `session.update` message.
9292

9393
> [!TIP]
94-
> For comprehensive descriptions of supported events and properties, see the [Azure OpenAI Realtime API events reference documentation](../openai/realtime-audio-reference.md?context=/azure/ai-services/speech-service/context/context). This document provides a reference for the event message properties that are enhancements via the voice live API.
94+
> For comprehensive descriptions of supported events and properties, see the [Azure OpenAI Realtime API events reference documentation](../openai/realtime-audio-reference.md?context=/azure/ai-services/speech-service/context/context). This document provides a reference for the event message properties that are enhancements via the Voice live API.
9595
9696
### Input audio properties
9797

@@ -125,11 +125,11 @@ Server echo cancellation enhances the input audio quality by removing the echo f
125125

126126
## Conversational enhancements
127127

128-
The voice live API offers conversational enhancements to provide robustness to the natural end-user conversation flow.
128+
The Voice live API offers conversational enhancements to provide robustness to the natural end-user conversation flow.
129129

130130
### Turn Detection Parameters
131131

132-
Turn detection is the process of detecting when the end-user started or stopped speaking. The voice live API builds on the Azure OpenAI Realtime API `turn_detection` property to configure turn detection. The `azure_semantic_vad` type and the advanced `end_of_utterance_detection` are key differentiators between the voice live API and the Azure OpenAI Realtime API.
132+
Turn detection is the process of detecting when the end-user started or stopped speaking. The Voice live API builds on the Azure OpenAI Realtime API `turn_detection` property to configure turn detection. The `azure_semantic_vad` type and the advanced `end_of_utterance_detection` are key differentiators between the Voice live API and the Azure OpenAI Realtime API.
133133

134134
| Property | Type | Required or optional | Description |
135135
|----------|----------|----------|------------|
@@ -139,7 +139,7 @@ Turn detection is the process of detecting when the end-user started or stopped
139139
| `speech_duration_ms` | integer | Optional | The duration of user's speech audio required to start detection. If not set or under 80 ms, the detector uses a default value of 80 ms. |
140140
| `silence_duration_ms` | integer | Optional | The duration of user's silence, measured in milliseconds, to detect the end of speech. |
141141
| `remove_filler_words` | boolean | Optional | Determines whether to remove filler words to reduce the false alarm rate. This property must be set to `true` when using `azure_semantic_vad`.<br/><br/>The default value is `false`. |
142-
| `end_of_utterance_detection` | object | Optional | Configuration for end of utterance detection. The voice live API offers advanced end-of-turn detection to indicate when the end-user stopped speaking while allowing for natural pauses. End of utterance detection can significantly reduce premature end-of-turn signals without adding user-perceivable latency. End of utterance detection can be used with either VAD selection.<br/><br/>Properties of `end_of_utterance_detection` include:<br/>-`model`: The model to use for end of utterance detection. The supported values are:<br/>&nbsp;&nbsp;`semantic_detection_v1` supporting English.<br/>&nbsp;&nbsp;`semantic_detection_v1_multilingual` supporting English, Spanish, French, Italian, German (DE), Japanese, Portuguese, Chinese, Korean, Hindi.<br/>Other languages will be bypassed.<br/>- `threshold`: Threshold to determine the end of utterance (0.0 to 1.0). The default value is 0.01.<br/>- `timeout`: Timeout in seconds. The default value is 2 seconds. <br/><br/>End of utterance detection currently doesn't support gpt-realtime, gpt-4o-mini-realtime, and phi4-mm-realtime.|
142+
| `end_of_utterance_detection` | object | Optional | Configuration for end of utterance detection. The Voice live API offers advanced end-of-turn detection to indicate when the end-user stopped speaking while allowing for natural pauses. End of utterance detection can significantly reduce premature end-of-turn signals without adding user-perceivable latency. End of utterance detection can be used with either VAD selection.<br/><br/>Properties of `end_of_utterance_detection` include:<br/>-`model`: The model to use for end of utterance detection. The supported values are:<br/>&nbsp;&nbsp;`semantic_detection_v1` supporting English.<br/>&nbsp;&nbsp;`semantic_detection_v1_multilingual` supporting English, Spanish, French, Italian, German (DE), Japanese, Portuguese, Chinese, Korean, Hindi.<br/>Other languages will be bypassed.<br/>- `threshold`: Threshold to determine the end of utterance (0.0 to 1.0). The default value is 0.01.<br/>- `timeout`: Timeout in seconds. The default value is 2 seconds. <br/><br/>End of utterance detection currently doesn't support gpt-realtime, gpt-4o-mini-realtime, and phi4-mm-realtime.|
143143

144144
Here's an example of end of utterance detection in a session object:
145145

articles/ai-services/speech-service/voice-live-language-support.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,15 +1,15 @@
11
---
22
title: Voice live API language support
33
titleSuffix: Azure AI services
4-
description: Learn about the languages supported by voice live API and how to configure them.
4+
description: Learn about the languages supported by Voice live API and how to configure them.
55
manager: nitinme
66
author: goergenj
77
ms.author: jagoerge
88
ms.service: azure-ai-speech
99
ms.topic: conceptual
10-
ms.date: 8/11/2025
10+
ms.date: 9/26/2025
1111
ms.custom: languages
12-
# Customer intent: As a developer, I want to learn about which languages are supported by the voice live API and how to configure them.
12+
# Customer intent: As a developer, I want to learn about which languages are supported by the Voice live API and how to configure them.
1313
---
1414

1515
# Voice live API supported languages (Preview)
@@ -18,7 +18,7 @@ ms.custom: languages
1818

1919
## Introduction
2020

21-
The voice live API supports multiple languages and configuration options. In this document, you learn which languages the voice live API supports and how to configure them.
21+
The Voice live API supports multiple languages and configuration options. In this document, you learn which languages the Voice live API supports and how to configure them.
2222

2323
## [Speech input](#tab/speechinput)
2424

@@ -207,6 +207,6 @@ If *Multilingual Voices* are used, the language output can optionally be control
207207

208208
## Related content
209209

210-
- Learn more about [How to use the voice live API](./voice-live-how-to.md)
211-
- Try out the [voice live API quickstart](./voice-live-quickstart.md)
210+
- Learn more about [How to use the Voice live API](./voice-live-how-to.md)
211+
- Try out the [Voice live API quickstart](./voice-live-quickstart.md)
212212
- See the [Voice live API reference](./voicelive-api-reference.md)

0 commit comments

Comments
 (0)