Skip to content

Commit 0abe86b

Browse files
authored
Merge pull request #4028 from eric-urban/eur/audio-models-april-1
audio models for AOAI
2 parents 8e2ffec + 520d1fa commit 0abe86b

27 files changed

+80
-61
lines changed

articles/ai-foundry/model-inference/concepts/content-filter.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ manager: nitinme
1414
# Content filtering for model inference in Azure AI services
1515

1616
> [!IMPORTANT]
17-
> The content filtering system isn't applied to prompts and completions processed by the Whisper model in Azure OpenAI. Learn more about the [Whisper model in Azure OpenAI](../../../ai-services/openai/concepts/models.md#whisper).
17+
> The content filtering system isn't applied to prompts and completions processed by the audio models such as Whisper in Azure OpenAI Service. Learn more about the [Audio API in Azure OpenAI](../../../ai-services/openai/concepts/models.md?tabs=audio#audio-models).
1818
1919
Azure AI model inference in Azure AI Services includes a content filtering system that works alongside core models and it's powered by [Azure AI Content Safety](https://azure.microsoft.com/products/cognitive-services/ai-content-safety). This system works by running both the prompt and completion through an ensemble of classification models designed to detect and prevent the output of harmful content. The content filtering system detects and takes action on specific categories of potentially harmful content in both input prompts and output completions. Variations in API configurations and application design might affect completions and thus filtering behavior.
2020

articles/ai-services/openai/api-version-deprecation.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -40,7 +40,7 @@ This version contains support for the latest Azure OpenAI features including:
4040
- [Text to speech](./text-to-speech-quickstart.md). [**Added in 2024-02-15-preview**]
4141
- [DALL-E 3](./dall-e-quickstart.md). [**Added in 2023-12-01-preview**]
4242
- [Fine-tuning](./how-to/fine-tuning.md). [**Added in 2023-10-01-preview**]
43-
- [Whisper](./whisper-quickstart.md). [**Added in 2023-09-01-preview**]
43+
- [Speech to text](./whisper-quickstart.md). [**Added in 2023-09-01-preview**]
4444
- [Function calling](./how-to/function-calling.md) [**Added in 2023-07-01-preview**]
4545
- [Retrieval augmented generation with your data feature](./use-your-data-quickstart.md). [**Added in 2023-06-01-preview**]
4646

articles/ai-services/openai/concepts/content-filter.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ manager: nitinme
1414
# Content filtering
1515

1616
> [!IMPORTANT]
17-
> The content filtering system isn't applied to prompts and completions processed by the Whisper model in Azure OpenAI Service. Learn more about the [Whisper model in Azure OpenAI](models.md#whisper).
17+
> The content filtering system isn't applied to prompts and completions processed by the audio models such as Whisper in Azure OpenAI Service. Learn more about the [Audio API in Azure OpenAI](models.md?tabs=audio#audio-models).
1818
1919
Azure OpenAI Service includes a content filtering system that works alongside core models, including DALL-E image generation models. This system works by running both the prompt and completion through an ensemble of classification models designed to detect and prevent the output of harmful content. The content filtering system detects and takes action on specific categories of potentially harmful content in both input prompts and output completions. Variations in API configurations and application design might affect completions and thus filtering behavior.
2020

articles/ai-services/openai/concepts/models.md

Lines changed: 23 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -27,8 +27,7 @@ Azure OpenAI Service is powered by a diverse set of models with different capabi
2727
| [GPT-3.5](#gpt-35) | A set of models that improve on GPT-3 and can understand and generate natural language and code. |
2828
| [Embeddings](#embeddings-models) | A set of models that can convert text into numerical vector form to facilitate text similarity. |
2929
| [DALL-E](#dall-e-models) | A series of models that can generate original images from natural language. |
30-
| [Whisper](#whisper-models) | A series of models in preview that can transcribe and translate speech to text. |
31-
| [Text to speech](#text-to-speech-models-preview) (Preview) | A series of models in preview that can synthesize text to speech. |
30+
| [Audio](#audio-models) | A series of models for speech to text, translation, and text to speech. |
3231

3332
## computer-use-preview
3433

@@ -236,17 +235,11 @@ OpenAI's MTEB benchmark testing found that even when the third generation model'
236235

237236
The DALL-E models generate images from text prompts that the user provides. DALL-E 3 is generally available for use with the REST APIs. DALL-E 2 and DALL-E 3 with client SDKs are in preview.
238237

239-
## Whisper
238+
## Audio API models
240239

241-
The Whisper models can be used for speech to text.
240+
The audio models via the `/audio` API can be used for speech to text, translation, and text to speech.
242241

243-
You can also use the Whisper model via Azure AI Speech [batch transcription](../../speech-service/batch-transcription-create.md) API. Check out [What is the Whisper model?](../../speech-service/whisper-overview.md) to learn more about when to use Azure AI Speech vs. Azure OpenAI Service.
244-
245-
## Text to speech (Preview)
246-
247-
The OpenAI text to speech models, currently in preview, can be used to synthesize text to speech.
248-
249-
You can also use the OpenAI text to speech voices via Azure AI Speech. To learn more, see [OpenAI text to speech voices via Azure OpenAI Service or via Azure AI Speech](../../speech-service/openai-voices.md#openai-text-to-speech-voices-via-azure-openai-service-or-via-azure-ai-speech) guide.
242+
For more information see [Audio models](#audio-models) in this article.
250243

251244
## Model summary table and region availability
252245

@@ -399,19 +392,31 @@ These models can only be used with Embedding API requests.
399392

400393
[!INCLUDE [Audio](../includes/model-matrix/standard-audio.md)]
401394

402-
### Whisper models
395+
### Speech to text models
403396

404-
| Model ID | Max Request (audio file size) |
405-
| --- | :---: |
406-
| `whisper` | 25 MB |
397+
| Model ID | Description | Max Request (audio file size) |
398+
| ----- | ----- | ----- |
399+
| `whisper` | General-purpose speech recognition model. | 25 MB |
400+
| `gpt-4o-transcribe` | Speech to text powered by GPT-4o. | 25 MB|
401+
| `gpt-4o-mini-transcribe` | Speech to text powered by GPT-4o mini. | 25 MB|
402+
403+
You can also use the Whisper model via Azure AI Speech [batch transcription](../../speech-service/batch-transcription-create.md) API. Check out [What is the Whisper model?](../../speech-service/whisper-overview.md) to learn more about when to use Azure AI Speech vs. Azure OpenAI Service.
404+
405+
### Speech translation models
406+
407+
| Model ID | Description | Max Request (audio file size) |
408+
| ----- | ----- | ----- |
409+
| `whisper` | General-purpose speech recognition model. | 25 MB |
407410

408411
### Text to speech models (Preview)
409412

410413
| Model ID | Description |
411414
| --- | :--- |
412-
| `tts` | The latest Azure OpenAI text to speech model, optimized for speed. |
413-
| `tts-hd` | The latest Azure OpenAI text to speech model, optimized for quality.|
414-
|
415+
| `tts` | Text to speech optimized for speed. |
416+
| `tts-hd` | Text to speech optimized for quality.|
417+
| `gpt-4o-mini-tts` | Text to speech model powered by GPT-4o mini. |
418+
419+
You can also use the OpenAI text to speech voices via Azure AI Speech. To learn more, see [OpenAI text to speech voices via Azure OpenAI Service or via Azure AI Speech](../../speech-service/openai-voices.md#openai-text-to-speech-voices-via-azure-openai-service-or-via-azure-ai-speech) guide.
415420

416421
# [Completions (Legacy)](#tab/standard-completions)
417422

articles/ai-services/openai/how-to/realtime-audio.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -116,7 +116,7 @@ Events can be sent and received in parallel and applications should generally ha
116116
Often, the first event sent by the caller on a newly established `/realtime` session is a [`session.update`](../realtime-audio-reference.md#realtimeclienteventsessionupdate) payload. This event controls a wide set of input and output behavior, with output and response generation properties then later overridable using the [`response.create`](../realtime-audio-reference.md#realtimeclienteventresponsecreate) event.
117117

118118
The [`session.update`](../realtime-audio-reference.md#realtimeclienteventsessionupdate) event can be used to configure the following aspects of the session:
119-
- Transcription of user input audio is opted into via the session's `input_audio_transcription` property. Specifying a transcription model (`whisper-1`) in this configuration enables the delivery of [`conversation.item.audio_transcription.completed`](../realtime-audio-reference.md#realtimeservereventconversationiteminputaudiotranscriptioncompleted) events.
119+
- Transcription of user input audio is opted into via the session's `input_audio_transcription` property. Specifying a transcription model (such as `whisper-1`) in this configuration enables the delivery of [`conversation.item.audio_transcription.completed`](../realtime-audio-reference.md#realtimeservereventconversationiteminputaudiotranscriptioncompleted) events.
120120
- Turn handling is controlled by the `turn_detection` property. This property's type can be set to `none` or `server_vad` as described in the [voice activity detection (VAD) and the audio buffer](#voice-activity-detection-vad-and-the-audio-buffer) section.
121121
- Tools can be configured to enable the server to call out to external services or functions to enrich the conversation. Tools are defined as part of the `tools` property in the session configuration.
122122

articles/ai-services/openai/includes/api-surface.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ Each API surface/specification encapsulates a different set of Azure OpenAI capa
2323
|:---|:----|:----|:----|:---|
2424
| **Control plane** | [`2024-06-01-preview`](/rest/api/aiservices/accountmanagement/operation-groups?view=rest-aiservices-accountmanagement-2024-06-01-preview&preserve-view=true) | [`2024-10-01`](/rest/api/aiservices/accountmanagement/deployments/create-or-update?view=rest-aiservices-accountmanagement-2024-10-01&tabs=HTTP&preserve-view=true) | [Spec files](https://github.com/Azure/azure-rest-api-specs/tree/main/specification/cognitiveservices/resource-manager/Microsoft.CognitiveServices) | Azure OpenAI shares a common control plane with all other Azure AI Services. The control plane API is used for things like [creating Azure OpenAI resources](/rest/api/aiservices/accountmanagement/accounts/create?view=rest-aiservices-accountmanagement-2023-05-01&tabs=HTTP&preserve-view=true), [model deployment](/rest/api/aiservices/accountmanagement/deployments/create-or-update?view=rest-aiservices-accountmanagement-2023-05-01&tabs=HTTP&preserve-view=true), and other higher level resource management tasks. The control plane also governs what is possible to do with capabilities like Azure Resource Manager, Bicep, Terraform, and Azure CLI.|
2525
| **Data plane - authoring** | `2025-03-01-preview` | `2024-10-21` | [Spec files](https://github.com/Azure/azure-rest-api-specs/tree/main/specification/cognitiveservices/data-plane/AzureOpenAI/authoring) | The data plane authoring API controls [fine-tuning](/rest/api/azureopenai/fine-tuning?view=rest-azureopenai-2024-08-01-preview&preserve-view=true), [file-upload](/rest/api/azureopenai/files/upload?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true), [ingestion jobs](/rest/api/azureopenai/ingestion-jobs/create?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true), [batch](/rest/api/azureopenai/batch?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true) and certain [model level queries](/rest/api/azureopenai/models/get?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true)
26-
| **Data plane - inference** | [`2025-03-01-preview`](/azure/ai-services/openai/reference-preview#data-plane-inference) | [`2024-10-21`](/azure/ai-services/openai/reference#data-plane-inference) | [Spec files](https://github.com/Azure/azure-rest-api-specs/tree/main/specification/cognitiveservices/data-plane/AzureOpenAI/inference) | The data plane inference API provides the inference capabilities/endpoints for features like completions, chat completions, embeddings, speech/whisper, on your data, Dall-e, assistants, etc. |
26+
| **Data plane - inference** | [`2025-03-01-preview`](/azure/ai-services/openai/reference-preview#data-plane-inference) | [`2024-10-21`](/azure/ai-services/openai/reference#data-plane-inference) | [Spec files](https://github.com/Azure/azure-rest-api-specs/tree/main/specification/cognitiveservices/data-plane/AzureOpenAI/inference) | The data plane inference API provides the inference capabilities/endpoints for features like completions, chat completions, embeddings, audio, on your data, Dall-e, assistants, etc. |
2727

2828
## Authentication
2929

articles/ai-services/openai/includes/api-versions/latest-inference.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -645,7 +645,7 @@ Transcribes audio into the input language.
645645
| Name | In | Required | Type | Description |
646646
|------|------|----------|------|-----------|
647647
| endpoint | path | Yes | string<br>url | Supported Azure OpenAI endpoints (protocol and hostname, for example: `https://aoairesource.openai.azure.com`. Replace "aoairesource" with your Azure OpenAI resource name). https://{your-resource-name}.openai.azure.com |
648-
| deployment-id | path | Yes | string | Deployment ID of the whisper model. |
648+
| deployment-id | path | Yes | string | Deployment ID of the speech to text model.<br/><br/>For information about supported models, see [/azure/ai-services/openai/concepts/models#audio-models]. |
649649
| api-version | query | Yes | string | API version |
650650

651651
### Request Header
@@ -731,7 +731,7 @@ Transcribes and translates input audio into English text.
731731
| Name | In | Required | Type | Description |
732732
|------|------|----------|------|-----------|
733733
| endpoint | path | Yes | string<br>url | Supported Azure OpenAI endpoints (protocol and hostname, for example: `https://aoairesource.openai.azure.com`. Replace "aoairesource" with your Azure OpenAI resource name). https://{your-resource-name}.openai.azure.com |
734-
| deployment-id | path | Yes | string | Deployment ID of the whisper model which was deployed. |
734+
| deployment-id | path | Yes | string | Deployment ID of the whisper model which was deployed.<br/><br/>For information about supported models, see [/azure/ai-services/openai/concepts/models#audio-models]. |
735735
| api-version | query | Yes | string | API version |
736736

737737
### Request Header
@@ -2318,6 +2318,6 @@ Completions extensions aren't part of the latest GA version of the Azure OpenAI
23182318

23192319
The Chat message object isn't part of the latest GA version of the Azure OpenAI data plane inference spec.
23202320

2321-
### Text to speech
2321+
### Text to speech (Preview)
23222322

23232323
Is not currently part of the latest Azure OpenAI GA version of the Azure OpenAI data plane inference spec. Refer to the latest [preview](../../reference-preview.md) version for this capability.

articles/ai-services/openai/includes/content-filter-configurability.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ recommendations: false
1111

1212

1313

14-
Azure OpenAI Service includes default safety settings applied to all models, excluding Azure OpenAI Whisper. These configurations provide you with a responsible experience by default, including content filtering models, blocklists, prompt transformation, [content credentials](../concepts/content-credentials.md), and others. [Read more about it here](/azure/ai-services/openai/concepts/default-safety-policies).
14+
Azure OpenAI Service includes default safety settings applied to all models, excluding audio API models such as Whisper. These configurations provide you with a responsible experience by default, including content filtering models, blocklists, prompt transformation, [content credentials](../concepts/content-credentials.md), and others. [Read more about it here](/azure/ai-services/openai/concepts/default-safety-policies).
1515

1616
All customers can also configure content filters and create custom safety policies that are tailored to their use case requirements. The configurability feature allows customers to adjust the settings, separately for prompts and completions, to filter content for each content category at different severity levels as described in the table below. Content detected at the 'safe' severity level is labeled in annotations but is not subject to filtering and isn't configurable.
1717

articles/ai-services/openai/includes/language-overview/go.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -190,12 +190,12 @@ import (
190190
)
191191

192192
func main() {
193-
azureOpenAIKey := os.Getenv("AOAI_WHISPER_API_KEY")
193+
azureOpenAIKey := os.Getenv("AOAI_AUDIO_API_KEY")
194194

195195
// Ex: "https://<your-azure-openai-host>.openai.azure.com"
196-
azureOpenAIEndpoint := os.Getenv("AOAI_WHISPER_ENDPOINT")
196+
azureOpenAIEndpoint := os.Getenv("AOAI_AUDIO_ENDPOINT")
197197

198-
modelDeploymentID := os.Getenv("AOAI_WHISPER_MODEL")
198+
modelDeploymentID := os.Getenv("AOAI_AUDIO_MODEL")
199199

200200
if azureOpenAIKey == "" || azureOpenAIEndpoint == "" || modelDeploymentID == "" {
201201
fmt.Fprintf(os.Stderr, "Skipping example, environment variables missing\n")

articles/ai-services/openai/includes/model-matrix/standard-audio.md

Lines changed: 10 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -9,13 +9,13 @@ ms.custom: references_regions
99
ms.date: 10/25/2024
1010
---
1111

12-
| **Region** | **tts**, **001** | **tts-hd**, **001** | **whisper**, **001** |
13-
|:-----------------|:----------------:|:-------------------:|:--------------------:|
14-
| eastus2 | - | - | |
15-
| northcentralus | | | |
16-
| norwayeast | - | - | |
17-
| southindia | - | - | |
18-
| swedencentral | | | |
19-
| switzerlandnorth | - | - | |
20-
| uaenorth | - | - | |
21-
| westeurope | - | - | |
12+
| **Region** | **tts**, **001** | **tts-hd**, **001** | **whisper**, **001** | **gpt-4o-mini-tts**, **001** | **gpt-4o-transcribe**, **001** | **gpt-4o-mini-transcribe **, **001** |
13+
|:-----------------|:----------------:|:-------------------:|:--------------------:|:--------------------:|:--------------------:|:--------------------:|
14+
| eastus2 | - | - || - || |
15+
| northcentralus |||||| |
16+
| norwayeast | - | - || - || |
17+
| southindia | - | - || - || |
18+
| swedencentral |||||| |
19+
| switzerlandnorth | - | - || - || |
20+
| uaenorth | - | - || - || |
21+
| westeurope | - | - || - || |

0 commit comments

Comments
 (0)