Skip to content

Commit df7d67e

Browse files
Merge pull request #4516 from MicrosoftDocs/main
Merged by Learn.Build PR Management system
2 parents 59073f6 + 37d2d71 commit df7d67e

File tree

9 files changed

+874
-68
lines changed

9 files changed

+874
-68
lines changed
Lines changed: 18 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -1,36 +1,39 @@
11
---
2-
title: "Azure AI Foundry docs: What's new for March 2025"
3-
description: "What's new in the Azure AI Foundry docs for March 2025."
2+
title: "Azure AI Foundry docs: What's new for April 2025"
3+
description: "What's new in the Azure AI Foundry docs for April 2025."
44
ms.author: smcdowell
55
author: skpmcdowell
66
ms.topic: whats-new
77
ms.subject: ai-studio
8-
ms.custom: March-2025
9-
ms.date: 04/02/2025
8+
ms.custom: April-2025
9+
ms.date: 05/03/2025
1010
---
1111

12-
# Azure AI Foundry docs: What's new for March 2025
12+
# Azure AI Foundry docs: What's new for April 2025
1313

14-
Welcome to what's new in the Azure AI Foundry docs for March 2025. This article lists some of the major changes to docs during this period.
14+
Welcome to what's new in the Azure AI Foundry docs for April 2025. This article lists some of the major changes to docs during this period.
1515

1616

1717
## Azure AI Foundry
1818

1919
### New articles
2020

21-
- [Featured models of Azure AI Foundry](../ai-foundry/concepts/models-featured.md)
22-
- [How to deploy NVIDIA Inference Microservices](../ai-foundry/how-to/deploy-nvidia-inference-microservice.md)
23-
- [How to use image and audio in chat completions with Azure AI model inference](../ai-foundry/model-inference/how-to/use-chat-multi-modal.md)
24-
- [Tutorial: Get started with DeepSeek-R1 reasoning model in Azure AI model inference](../ai-foundry/model-inference/tutorials/get-started-deepseek-r1.md)
21+
- [AI Red Teaming Agent (preview)](../ai-foundry/concepts/ai-red-teaming-agent.md)
22+
- [Evaluate your AI agents locally with Azure AI Evaluation SDK (preview)](../ai-foundry/how-to/develop/agent-evaluate-sdk.md)
23+
- [How to use structured outputs for chat models](../ai-foundry/model-inference/how-to/use-structured-outputs.md)
24+
- [Run automated safety scans with AI Red Teaming Agent (preview)](../ai-foundry/how-to/develop/run-scans-ai-red-teaming-agent.md)
25+
- [Work with Azure AI Agent Service in Visual Studio Code (Preview)](../ai-foundry/how-to/develop/vs-code-agents.md)
26+
- [Work with the Azure AI Foundry for Visual Studio Code extension (Preview)](../ai-foundry/how-to/develop/get-started-projects-vs-code.md)
2527

2628

2729
### Updated articles
2830

29-
- [Deploy a flow for real-time inference](../ai-foundry/how-to/flow-deploy.md)
30-
- [Fine-tune models using serverless APIs in Azure AI Foundry](../ai-foundry/how-to/fine-tune-serverless.md)
31-
- [How to deploy and inference a managed compute deployment with code](../ai-foundry/how-to/deploy-models-managed.md)
32-
- [How to trace your application with Azure AI Foundry project library](../ai-foundry/how-to/develop/trace-local-sdk.md)
33-
- [Monitor quality and token usage of deployed prompt flow applications](../ai-foundry/how-to/monitor-quality-safety.md)
31+
- [Evaluate your AI agents locally with Azure AI Evaluation SDK (preview)](../ai-foundry/how-to/develop/agent-evaluate-sdk.md)
32+
- [Evaluate your Generative AI application locally with the Azure AI Evaluation SDK](../ai-foundry/how-to/develop/evaluate-sdk.md)
33+
- [Evaluation and monitoring metrics for generative AI](../ai-foundry/concepts/evaluation-metrics-built-in.md)
34+
- [Fine-tune models using serverless APIs in Azure AI Foundry](../ai-foundry/how-to/fine-tune-serverless.md)
35+
- [How to configure a private link for Azure AI Foundry hubs](../ai-foundry/how-to/configure-private-link.md)
36+
- [How to use MedImageParse healthcare AI models for segmentation of medical images](../ai-foundry/how-to/healthcare-ai/deploy-medimageparse.md)
3437

3538

3639

articles/ai-services/openai/how-to/realtime-audio.md

Lines changed: 12 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -53,7 +53,7 @@ Often, the first event sent by the caller on a newly established `/realtime` ses
5353

5454
The [`session.update`](../realtime-audio-reference.md#realtimeclienteventsessionupdate) event can be used to configure the following aspects of the session:
5555
- Transcription of user input audio is opted into via the session's `input_audio_transcription` property. Specifying a transcription model (such as `whisper-1`) in this configuration enables the delivery of [`conversation.item.audio_transcription.completed`](../realtime-audio-reference.md#realtimeservereventconversationiteminputaudiotranscriptioncompleted) events.
56-
- Turn handling is controlled by the `turn_detection` property. This property's type can be set to `none` or `server_vad` as described in the [voice activity detection (VAD) and the audio buffer](#voice-activity-detection-vad-and-the-audio-buffer) section.
56+
- Turn handling is controlled by the `turn_detection` property. This property's type can be set to `none`, `semantic_vad`, or `server_vad` as described in the [voice activity detection (VAD) and the audio buffer](#voice-activity-detection-vad-and-the-audio-buffer) section.
5757
- Tools can be configured to enable the server to call out to external services or functions to enrich the conversation. Tools are defined as part of the `tools` property in the session configuration.
5858

5959
An example `session.update` that configures several aspects of the session, including tools, follows. All session parameters are optional and can be omitted if not needed.
@@ -144,9 +144,12 @@ You can also construct a custom context that the model uses outside of the sessi
144144

145145
The server maintains an input audio buffer containing client-provided audio that hasn't yet been committed to the conversation state.
146146

147-
One of the key [session-wide](#session-configuration) settings is `turn_detection`, which controls how data flow is handled between the caller and model. The `turn_detection` setting can be set to `none` or `server_vad` (to use [server-side voice activity detection](#server-decision-mode)).
147+
One of the key [session-wide](#session-configuration) settings is `turn_detection`, which controls how data flow is handled between the caller and model. The `turn_detection` setting can be set to `none`, `semantic_vad`, or `server_vad` (to use [server-side voice activity detection](#server-decision-mode)).
148148

149-
By default, voice activity detection (VAD) is enabled, and the server automatically generates responses when it detects the end of speech in the input audio buffer. You can change the behavior by setting the `turn_detection` property in the session configuration.
149+
- `server_vad`: Automatically chunks the audio based on periods of silence.
150+
- `semantic_vad`: Chunks the audio when the model believes based on the words said by the user that they have completed their utterance.
151+
152+
By default, server VAD (`server_vad`) is enabled, and the server automatically generates responses when it detects the end of speech in the input audio buffer. You can change the behavior by setting the `turn_detection` property in the session configuration.
150153

151154
### Without server decision mode
152155

@@ -201,6 +204,12 @@ sequenceDiagram
201204
Server->>Client: conversation.item.created
202205
-->
203206

207+
### Semantic VAD
208+
209+
Semantic VAD detects when the user has finished speaking based on the words they have uttered. The input audio is scored based on the probability that the user is done speaking. When the probability is low the model will wait for a timeout. When the probability is high there's no need to wait.
210+
211+
With the (`semantic_vad`) mode, the model is less likely to interrupt the user during a speech-to-speech conversation, or chunk a transcript before the user is done speaking.
212+
204213
### VAD without automatic response generation
205214

206215
You can use server-side voice activity detection (VAD) without automatic response generation. This approach can be useful when you want to implement some degree of moderation.

0 commit comments

Comments
 (0)