You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
title: "Azure AI Foundry docs: What's new for March 2025"
3
-
description: "What's new in the Azure AI Foundry docs for March 2025."
2
+
title: "Azure AI Foundry docs: What's new for April 2025"
3
+
description: "What's new in the Azure AI Foundry docs for April 2025."
4
4
ms.author: smcdowell
5
5
author: skpmcdowell
6
6
ms.topic: whats-new
7
7
ms.subject: ai-studio
8
-
ms.custom: March-2025
9
-
ms.date: 04/02/2025
8
+
ms.custom: April-2025
9
+
ms.date: 05/03/2025
10
10
---
11
11
12
-
# Azure AI Foundry docs: What's new for March 2025
12
+
# Azure AI Foundry docs: What's new for April 2025
13
13
14
-
Welcome to what's new in the Azure AI Foundry docs for March 2025. This article lists some of the major changes to docs during this period.
14
+
Welcome to what's new in the Azure AI Foundry docs for April 2025. This article lists some of the major changes to docs during this period.
15
15
16
16
17
17
## Azure AI Foundry
18
18
19
19
### New articles
20
20
21
-
-[Featured models of Azure AI Foundry](../ai-foundry/concepts/models-featured.md)
22
-
-[How to deploy NVIDIA Inference Microservices](../ai-foundry/how-to/deploy-nvidia-inference-microservice.md)
23
-
-[How to use image and audio in chat completions with Azure AI model inference](../ai-foundry/model-inference/how-to/use-chat-multi-modal.md)
24
-
-[Tutorial: Get started with DeepSeek-R1 reasoning model in Azure AI model inference](../ai-foundry/model-inference/tutorials/get-started-deepseek-r1.md)
21
+
-[AI Red Teaming Agent (preview)](../ai-foundry/concepts/ai-red-teaming-agent.md)
22
+
-[Evaluate your AI agents locally with Azure AI Evaluation SDK (preview)](../ai-foundry/how-to/develop/agent-evaluate-sdk.md)
23
+
-[How to use structured outputs for chat models](../ai-foundry/model-inference/how-to/use-structured-outputs.md)
24
+
-[Run automated safety scans with AI Red Teaming Agent (preview)](../ai-foundry/how-to/develop/run-scans-ai-red-teaming-agent.md)
25
+
-[Work with Azure AI Agent Service in Visual Studio Code (Preview)](../ai-foundry/how-to/develop/vs-code-agents.md)
26
+
-[Work with the Azure AI Foundry for Visual Studio Code extension (Preview)](../ai-foundry/how-to/develop/get-started-projects-vs-code.md)
25
27
26
28
27
29
### Updated articles
28
30
29
-
-[Deploy a flow for real-time inference](../ai-foundry/how-to/flow-deploy.md)
30
-
-[Fine-tune models using serverless APIs in Azure AI Foundry](../ai-foundry/how-to/fine-tune-serverless.md)
31
-
-[How to deploy and inference a managed compute deployment with code](../ai-foundry/how-to/deploy-models-managed.md)
32
-
-[How to trace your application with Azure AI Foundry project library](../ai-foundry/how-to/develop/trace-local-sdk.md)
33
-
-[Monitor quality and token usage of deployed prompt flow applications](../ai-foundry/how-to/monitor-quality-safety.md)
31
+
-[Evaluate your AI agents locally with Azure AI Evaluation SDK (preview)](../ai-foundry/how-to/develop/agent-evaluate-sdk.md)
32
+
-[Evaluate your Generative AI application locally with the Azure AI Evaluation SDK](../ai-foundry/how-to/develop/evaluate-sdk.md)
33
+
-[Evaluation and monitoring metrics for generative AI](../ai-foundry/concepts/evaluation-metrics-built-in.md)
34
+
-[Fine-tune models using serverless APIs in Azure AI Foundry](../ai-foundry/how-to/fine-tune-serverless.md)
35
+
-[How to configure a private link for Azure AI Foundry hubs](../ai-foundry/how-to/configure-private-link.md)
36
+
-[How to use MedImageParse healthcare AI models for segmentation of medical images](../ai-foundry/how-to/healthcare-ai/deploy-medimageparse.md)
Copy file name to clipboardExpand all lines: articles/ai-services/openai/how-to/realtime-audio.md
+12-3Lines changed: 12 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -53,7 +53,7 @@ Often, the first event sent by the caller on a newly established `/realtime` ses
53
53
54
54
The [`session.update`](../realtime-audio-reference.md#realtimeclienteventsessionupdate) event can be used to configure the following aspects of the session:
55
55
- Transcription of user input audio is opted into via the session's `input_audio_transcription` property. Specifying a transcription model (such as `whisper-1`) in this configuration enables the delivery of [`conversation.item.audio_transcription.completed`](../realtime-audio-reference.md#realtimeservereventconversationiteminputaudiotranscriptioncompleted) events.
56
-
- Turn handling is controlled by the `turn_detection` property. This property's type can be set to `none` or `server_vad` as described in the [voice activity detection (VAD) and the audio buffer](#voice-activity-detection-vad-and-the-audio-buffer) section.
56
+
- Turn handling is controlled by the `turn_detection` property. This property's type can be set to `none`, `semantic_vad`, or `server_vad` as described in the [voice activity detection (VAD) and the audio buffer](#voice-activity-detection-vad-and-the-audio-buffer) section.
57
57
- Tools can be configured to enable the server to call out to external services or functions to enrich the conversation. Tools are defined as part of the `tools` property in the session configuration.
58
58
59
59
An example `session.update` that configures several aspects of the session, including tools, follows. All session parameters are optional and can be omitted if not needed.
@@ -144,9 +144,12 @@ You can also construct a custom context that the model uses outside of the sessi
144
144
145
145
The server maintains an input audio buffer containing client-provided audio that hasn't yet been committed to the conversation state.
146
146
147
-
One of the key [session-wide](#session-configuration) settings is `turn_detection`, which controls how data flow is handled between the caller and model. The `turn_detection` setting can be set to `none` or `server_vad` (to use [server-side voice activity detection](#server-decision-mode)).
147
+
One of the key [session-wide](#session-configuration) settings is `turn_detection`, which controls how data flow is handled between the caller and model. The `turn_detection` setting can be set to `none`, `semantic_vad`, or `server_vad` (to use [server-side voice activity detection](#server-decision-mode)).
148
148
149
-
By default, voice activity detection (VAD) is enabled, and the server automatically generates responses when it detects the end of speech in the input audio buffer. You can change the behavior by setting the `turn_detection` property in the session configuration.
149
+
-`server_vad`: Automatically chunks the audio based on periods of silence.
150
+
-`semantic_vad`: Chunks the audio when the model believes based on the words said by the user that they have completed their utterance.
151
+
152
+
By default, server VAD (`server_vad`) is enabled, and the server automatically generates responses when it detects the end of speech in the input audio buffer. You can change the behavior by setting the `turn_detection` property in the session configuration.
150
153
151
154
### Without server decision mode
152
155
@@ -201,6 +204,12 @@ sequenceDiagram
201
204
Server->>Client: conversation.item.created
202
205
-->
203
206
207
+
### Semantic VAD
208
+
209
+
Semantic VAD detects when the user has finished speaking based on the words they have uttered. The input audio is scored based on the probability that the user is done speaking. When the probability is low the model will wait for a timeout. When the probability is high there's no need to wait.
210
+
211
+
With the (`semantic_vad`) mode, the model is less likely to interrupt the user during a speech-to-speech conversation, or chunk a transcript before the user is done speaking.
212
+
204
213
### VAD without automatic response generation
205
214
206
215
You can use server-side voice activity detection (VAD) without automatic response generation. This approach can be useful when you want to implement some degree of moderation.
0 commit comments