You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-services/openai/how-to/audio-real-time.md
+4-4Lines changed: 4 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -12,11 +12,11 @@ ms.custom: references_regions
12
12
recommendations: false
13
13
---
14
14
15
-
# GPT-4o Realtime API for speech and audio
15
+
# GPT-4o Realtime API for speech and audio (Preview)
16
16
17
17
Azure OpenAI GPT-4o Realtime API for speech and audio is part of the GPT-4o model family that supports low-latency, "speech in, speech out" conversational interactions. The GPT-4o audio `realtime` API is designed to handle real-time, low-latency conversational interactions, making it a great fit for use cases involving live interactions between a user and a model, such as customer support agents, voice assistants, and real-time translators.
18
18
19
-
Most users of this API need to deliver and receive audio from an end-user in real time, including applications that use WebRTC or a telephony system. The Realtime API isn't designed to connect directly to end user devices and relies on client integrations to terminate end user audio streams.
19
+
Most users of the Realtime API need to deliver and receive audio from an end-user in real time, including applications that use WebRTC or a telephony system. The Realtime API isn't designed to connect directly to end user devices and relies on client integrations to terminate end user audio streams.
20
20
21
21
## Supported models
22
22
@@ -43,7 +43,7 @@ Support for the Realtime API was first added in API version `2024-10-01-preview`
43
43
44
44
Before you can use GPT-4o real-time audio, you need a deployment of the `gpt-4o-realtime-preview` model in a supported region as described in the [supported models](#supported-models) section.
45
45
46
-
You can deploy the model from the Azure OpenAI model catalog or from your project in AI Studio. Follow these steps to deploy a `gpt-4o-realtime-preview` model from the [AI Studio model catalog](../../../ai-studio/how-to/model-catalog-overview.md):
46
+
You can deploy the model from the [Azure AI Studio model catalog](../../../ai-studio/how-to/model-catalog-overview.md) or from your project in AI Studio. Follow these steps to deploy a `gpt-4o-realtime-preview` model from the model catalog:
47
47
48
48
1. Sign in to [AI Studio](https://ai.azure.com) and go to the **Home** page.
49
49
1. Select **Model catalog** from the left sidebar.
@@ -54,7 +54,7 @@ You can deploy the model from the Azure OpenAI model catalog or from your projec
54
54
1. Modify other default settings depending on your requirements.
55
55
1. Select **Deploy**. You land on the deployment details page.
56
56
57
-
Now that you have a deployment of the `gpt-4o-realtime-preview` model, you can use the playground to interact with the model in real time. Select **Early access playground** from the list of playgrounds in the left pane.
57
+
Now that you have a deployment of the `gpt-4o-realtime-preview` model, you can use the Realtime API to interact with it in real time.
Copy file name to clipboardExpand all lines: articles/ai-services/openai/whats-new.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -38,7 +38,7 @@ The `max_tokens` parameter has been deprecated and replaced with the new `max_co
38
38
39
39
Models are available for standard and global standard deployment in East US2 and Sweden Central for approved customers.
40
40
41
-
### New GPT-4o real-time audio public preview
41
+
### New GPT-4o Realtime API for speech and audio public preview
42
42
43
43
Azure OpenAI GPT-4o audio is part of the GPT-4o model family that supports low-latency, "speech in, speech out" conversational interactions. The GPT-4o audio `realtime` API is designed to handle real-time, low-latency conversational interactions, making it a great fit for use cases involving live interactions between a user and a model, such as customer support agents, voice assistants, and real-time translators.
0 commit comments