Skip to content

Commit 5bad034

Browse files
committed
Merge branch 'main' into release-azure-search
2 parents 2eb67b9 + d4aaed2 commit 5bad034

17 files changed

+573
-539
lines changed

articles/ai-services/openai/api-version-deprecation.md

Lines changed: 9 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ services: cognitive-services
55
manager: nitinme
66
ms.service: azure-ai-openai
77
ms.topic: conceptual
8-
ms.date: 09/05/2024
8+
ms.date: 10/01/2024
99
author: mrbullwinkle
1010
ms.author: mbullwin
1111
recommendations: false
@@ -23,7 +23,7 @@ This article is to help you understand the support lifecycle for the Azure OpenA
2323

2424
Azure OpenAI API latest release:
2525

26-
- Inference: [2024-08-01-preview](https://github.com/Azure/azure-rest-api-specs/blob/main/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2024-08-01-preview/inference.json)
26+
- Inference: [2024-09-01-preview](https://github.com/Azure/azure-rest-api-specs/blob/main/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2024-09-01-preview/inference.json)
2727
- Authoring: [2024-08-01-preview](https://github.com/Azure/azure-rest-api-specs/blob/main/specification/cognitiveservices/data-plane/AzureOpenAI/authoring/preview/2024-08-01-preview/azureopenai.json)
2828

2929
This version contains support for the latest Azure OpenAI features including:
@@ -38,6 +38,13 @@ This version contains support for the latest Azure OpenAI features including:
3838
- [Function calling](./how-to/function-calling.md) [**Added in 2023-07-01-preview**]
3939
- [Retrieval augmented generation with your data feature](./use-your-data-quickstart.md). [**Added in 2023-06-01-preview**]
4040

41+
## Changes between 2024-09-01-preview and 2024-08-01-preview
42+
43+
- `max_completion_tokens` added to support `o1-preview` and `o1-mini` models. `max_tokens` does not work with the **o1 series** models.
44+
- `parallel_tool_calls` added.
45+
- `completion_tokens_details` & `reasoning_tokens` added.
46+
- `stream_options` & `include_usage` added.
47+
4148
## Changes between 2024-07-01-preview and 2024-08-01-preview API specification
4249

4350
- [Structured outputs support](./how-to/structured-outputs.md).

articles/ai-services/openai/how-to/audio-real-time.md

Lines changed: 29 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -5,10 +5,11 @@ description: Learn how to use GPT-4o Realtime API for speech and audio with Azur
55
manager: nitinme
66
ms.service: azure-ai-openai
77
ms.topic: how-to
8-
ms.date: 10/1/2024
8+
ms.date: 10/3/2024
99
author: eric-urban
1010
ms.author: eur
1111
ms.custom: references_regions
12+
zone_pivot_groups: openai-studio-js
1213
recommendations: false
1314
---
1415

@@ -54,14 +55,35 @@ You can deploy the model from the [Azure AI Studio model catalog](../../../ai-st
5455
1. Modify other default settings depending on your requirements.
5556
1. Select **Deploy**. You land on the deployment details page.
5657

57-
Now that you have a deployment of the `gpt-4o-realtime-preview` model, you can use the Realtime API to interact with it in real time.
58+
Now that you have a deployment of the `gpt-4o-realtime-preview` model, you can use the AI Studio **Real-time audio** playground or Realtime API to interact with it in real time.
5859

59-
## Use the GPT-4o Realtime API
60+
## Use the GPT-4o real-time audio
6061

6162
> [!TIP]
62-
> A playground for GPT-4o real-time audio is coming soon to [Azure AI Studio](https://ai.azure.com). You can already use the API directly in your application.
63+
> Right now, the fastest way to get started development with the GPT-4o Realtime API is to download the sample code from the [Azure OpenAI GPT-4o real-time audio repository on GitHub](https://github.com/azure-samples/aoai-realtime-audio-sdk).
6364
64-
Right now, the fastest way to get started with the GPT-4o Realtime API is to download the sample code from the [Azure OpenAI GPT-4o real-time audio repository on GitHub](https://github.com/azure-samples/aoai-realtime-audio-sdk).
65+
::: zone pivot="programming-language-ai-studio"
66+
67+
To chat with your deployed `gpt-4o-realtime-preview` model in the [Azure AI Studio](https://ai.azure.com) **Real-time audio** playground, follow these steps:
68+
69+
1. Go to your project in [Azure AI Studio](https://ai.azure.com).
70+
1. Select **Playgrounds** > **Real-time audio** from the left pane.
71+
1. Select your deployed `gpt-4o-realtime-preview` model from the **Deployment** dropdown.
72+
1. Select **Enable microphone** to allow the browser to access your microphone. If you already granted permission, you can skip this step.
73+
74+
:::image type="content" source="../media/how-to/real-time/real-time-playground.png" alt-text="Screenshot of the real-time audio playground with the deployed model selected." lightbox="../media/how-to/real-time/real-time-playground.png":::
75+
76+
1. Optionally you can edit contents in the **Give the model instructions and context** text box. Give the model instructions about how it should behave and any context it should reference when generating a response. You can describe the assistant's personality, tell it what it should and shouldn't answer, and tell it how to format responses.
77+
1. Optionally, change settings such as threshold, prefix padding, and silence duration.
78+
1. Select **Start listening** to start the session. You can speak into the microphone to start a chat.
79+
80+
:::image type="content" source="../media/how-to/real-time/real-time-playground-start-listening.png" alt-text="Screenshot of the real-time audio playground with the start listening button and microphone access enabled." lightbox="../media/how-to/real-time/real-time-playground-start-listening.png":::
81+
82+
1. You can interrupt the chat at any time by speaking. You can end the chat by selecting the **Stop listening** button.
83+
84+
::: zone-end
85+
86+
::: zone pivot="programming-language-javascript"
6587

6688
The JavaScript web sample demonstrates how to use the GPT-4o Realtime API to interact with the model in real time. The sample code includes a simple web interface that captures audio from the user's microphone and sends it to the model for processing. The model responds with text and audio, which the sample code renders in the web interface.
6789

@@ -103,6 +125,8 @@ You can run the sample code locally on your machine by following these steps. Re
103125
1. You should see a `<< Session Started >>` message in the main output. Then you can speak into the microphone to start a chat.
104126
1. You can interrupt the chat at any time by speaking. You can end the chat by selecting the **Stop** button.
105127
128+
::: zone-end
129+
106130
## Related content
107131
108132
* Learn more about Azure OpenAI [deployment types](./deployment-types.md)

articles/ai-services/openai/includes/api-surface.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -22,8 +22,8 @@ Each API surface/specification encapsulates a different set of Azure OpenAI capa
2222
| API | Latest preview release | Latest GA release | Specifications | Description |
2323
|:---|:----|:----|:----|:---|
2424
| **Control plane** | [`2024-06-01-preview`](/rest/api/aiservices/accountmanagement/operation-groups?view=rest-aiservices-accountmanagement-2024-06-01-preview&preserve-view=true) | [`2023-05-01`](/rest/api/aiservices/accountmanagement/deployments/create-or-update?view=rest-aiservices-accountmanagement-2023-05-01&tabs=HTTP&preserve-view=true) | [Spec files](https://github.com/Azure/azure-rest-api-specs/tree/main/specification/cognitiveservices/resource-manager/Microsoft.CognitiveServices) | Azure OpenAI shares a common control plane with all other Azure AI Services. The control plane API is used for things like [creating Azure OpenAI resources](/rest/api/aiservices/accountmanagement/accounts/create?view=rest-aiservices-accountmanagement-2023-05-01&tabs=HTTP&preserve-view=true), [model deployment](/rest/api/aiservices/accountmanagement/deployments/create-or-update?view=rest-aiservices-accountmanagement-2023-05-01&tabs=HTTP&preserve-view=true), and other higher level resource management tasks. The control plane also governs what is possible to do with capabilities like Azure Resource Manager, Bicep, Terraform, and Azure CLI.|
25-
| **Data plane - authoring** | `2024-08-01-preview` | [`2024-06-01`](/rest/api/azureopenai/operation-groups?view=rest-azureopenai-2024-06-01&preserve-view=true) | [Spec files](https://github.com/Azure/azure-rest-api-specs/tree/main/specification/cognitiveservices/data-plane/AzureOpenAI/authoring) | The data plane authoring API controls [fine-tuning](/rest/api/azureopenai/fine-tuning?view=rest-azureopenai-2024-08-01-preview&preserve-view=true), [file-upload](/rest/api/azureopenai/files/upload?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true), [ingestion jobs](/rest/api/azureopenai/ingestion-jobs/create?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true), [batch](/rest/api/azureopenai/batch?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true) and certain [model level queries](/rest/api/azureopenai/models/get?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true)
26-
| **Data plane - inference** | [`2024-08-01-preview`](/azure/ai-services/openai/reference-preview#data-plane-inference) | [`2024-06-01`](/azure/ai-services/openai/reference#data-plane-inference) | [Spec files](https://github.com/Azure/azure-rest-api-specs/tree/main/specification/cognitiveservices/data-plane/AzureOpenAI/inference) | The data plane inference API provides the inference capabilities/endpoints for features like completions, chat completions, embeddings, speech/whisper, on your data, Dall-e, assistants, etc. |
25+
| **Data plane - authoring** | [`2024-08-01-preview`](/rest/api/azureopenai/operation-groups?view=rest-azureopenai-2024-08-01-preview&preserve-view=true) | [`2024-06-01`](/rest/api/azureopenai/operation-groups?view=rest-azureopenai-2024-06-01&preserve-view=true) | [Spec files](https://github.com/Azure/azure-rest-api-specs/tree/main/specification/cognitiveservices/data-plane/AzureOpenAI/authoring) | The data plane authoring API controls [fine-tuning](/rest/api/azureopenai/fine-tuning?view=rest-azureopenai-2024-08-01-preview&preserve-view=true), [file-upload](/rest/api/azureopenai/files/upload?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true), [ingestion jobs](/rest/api/azureopenai/ingestion-jobs/create?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true), [batch](/rest/api/azureopenai/batch?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true) and certain [model level queries](/rest/api/azureopenai/models/get?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true)
26+
| **Data plane - inference** | [`2024-09-01-preview`](/azure/ai-services/openai/reference-preview#data-plane-inference) | [`2024-06-01`](/azure/ai-services/openai/reference#data-plane-inference) | [Spec files](https://github.com/Azure/azure-rest-api-specs/tree/main/specification/cognitiveservices/data-plane/AzureOpenAI/inference) | The data plane inference API provides the inference capabilities/endpoints for features like completions, chat completions, embeddings, speech/whisper, on your data, Dall-e, assistants, etc. |
2727

2828
## Authentication
2929

0 commit comments

Comments
 (0)