Skip to content

Commit 9251fbb

Browse files
authored
Merge branch 'MicrosoftDocs:main' into heidist-fresh
2 parents d78a13a + 00a4cc4 commit 9251fbb

17 files changed

+246
-199
lines changed

articles/ai-services/document-intelligence/containers/install-run.md

Lines changed: 12 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -5,10 +5,8 @@ description: Use the Docker containers for Document Intelligence on-premises to
55
author: laujan
66
manager: nitinme
77
ms.service: azure-ai-document-intelligence
8-
ms.custom:
9-
- ignite-2023
108
ms.topic: how-to
11-
ms.date: 07/09/2024
9+
ms.date: 10/01/2024
1210
ms.author: lajanuar
1311
---
1412

@@ -18,30 +16,27 @@ ms.author: lajanuar
1816
<!-- markdownlint-disable MD024 -->
1917
<!-- markdownlint-disable MD051 -->
2018

21-
:::moniker range="doc-intel-2.1.0 || doc-intel-4.0.0"
22-
Support for containers is currently available with Document Intelligence version `2022-08-31 (GA)` for all models and `2023-07-31 (GA)` for Read, Layout, ID Document, Receipt, and Invoice models:
23-
24-
* [REST API `2022-08-31 (GA)`](/rest/api/aiservices/document-models/analyze-document?view=rest-aiservices-v3.0%20(2022-08-31)&preserve-view=true&tabs=HTTP)
25-
* [REST API `2023-07-31 (GA)`](/rest/api/aiservices/document-models/analyze-document?view=rest-aiservices-v3.1%20(2023-07-31)&tabs=HTTP&preserve-view=true)
26-
* [Client libraries targeting `REST API 2022-08-31 (GA)`](../sdk-overview-v3-0.md)
27-
* [Client libraries targeting `REST API 2023-07-31 (GA)`](../sdk-overview-v3-1.md)
28-
29-
✔️ See [**Install and run Document Intelligence containers**](?view=doc-intel-3.1.0&preserve-view=true) for supported container documentation.
30-
31-
:::moniker-end
32-
33-
:::moniker range="doc-intel-3.0.0 || doc-intel-3.1.0"
19+
:::moniker range=">=doc-intel-2.1.0"
3420

3521
**This content applies to:** ![checkmark](../media/yes-icon.png) **v3.0 (GA)** ![checkmark](../media/yes-icon.png) **v3.1 (GA)**
3622

3723
Azure AI Document Intelligence is an Azure AI service that lets you build automated data processing software using machine-learning technology. Document Intelligence enables you to identify and extract text, key/value pairs, selection marks, table data, and more from your documents. The results are delivered as structured data that ../includes the relationships in the original file.
3824

39-
In this article you learn how to download, install, and run Document Intelligence containers. Containers enable you to run the Document Intelligence service in your own environment. Containers are great for specific security and data governance requirements.
25+
In this article you can learn how to download, install, and run Document Intelligence containers. Containers enable you to run the Document Intelligence service in your own environment. Containers are great for specific security and data governance requirements.
4026

4127
* **Read**, **Layout**, **ID Document**, **Receipt**, and **Invoice** models are supported by Document Intelligence v3.1 containers.
4228

4329
* **Read**, **Layout**, **General Document**, **Business Card**, and **Custom** models are supported by Document Intelligence v3.0 containers.
4430

31+
## Version support
32+
33+
Support for containers is currently available with Document Intelligence version `v3.0: 2022-08-31 (GA)` for all models and `v3.1 2023-07-31 (GA)` for Read, Layout, ID Document, Receipt, and Invoice models:
34+
35+
* [REST API `v3.0: 2022-08-31 (GA)`](/rest/api/aiservices/document-models/analyze-document?view=rest-aiservices-v3.0%20(2022-08-31)&preserve-view=true&tabs=HTTP)
36+
* [REST API `v3.1: 2023-07-31 (GA)`](/rest/api/aiservices/document-models/analyze-document?view=rest-aiservices-v3.1%20(2023-07-31)&tabs=HTTP&preserve-view=true)
37+
* [Client libraries targeting `REST API v3.0: 2022-08-31 (GA)`](../sdk-overview-v3-0.md)
38+
* [Client libraries targeting `REST API v3.1: 2023-07-31 (GA)`](../sdk-overview-v3-1.md)
39+
4540
## Prerequisites
4641

4742
To get started, you need an active [**Azure account**](https://azure.microsoft.com/free/cognitive-services/). If you don't have one, you can [**create a free account**](https://azure.microsoft.com/free/).

articles/ai-services/openai/concepts/content-filter.md

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -83,8 +83,6 @@ Detecting indirect attacks requires using document delimiters when constructing
8383

8484
[!INCLUDE [content-filter-configurability](../includes/content-filter-configurability.md)]
8585

86-
87-
8886
## Scenario details
8987

9088
When the content filtering system detects harmful content, you receive either an error on the API call if the prompt was deemed inappropriate, or the `finish_reason` on the response will be `content_filter` to signify that some of the completion was filtered. When building your application or system, you'll want to account for these scenarios where the content returned by the Completions API is filtered, which might result in content that is incomplete. How you act on this information will be application specific. The behavior can be summarized in the following points:
@@ -1012,4 +1010,4 @@ As part of your application design, consider the following best practices to del
10121010
- Apply for modified content filters via [this form](https://ncv.microsoft.com/uEfCgnITdR).
10131011
- Azure OpenAI content filtering is powered by [Azure AI Content Safety](https://azure.microsoft.com/products/cognitive-services/ai-content-safety).
10141012
- Learn more about understanding and mitigating risks associated with your application: [Overview of Responsible AI practices for Azure OpenAI models](/legal/cognitive-services/openai/overview?context=/azure/ai-services/openai/context/context).
1015-
- Learn more about how data is processed in connection with content filtering and abuse monitoring: [Data, privacy, and security for Azure OpenAI Service](/legal/cognitive-services/openai/data-privacy?context=/azure/ai-services/openai/context/context#preventing-abuse-and-harmful-content-generation).
1013+
- Learn more about how data is processed in connection with content filtering and abuse monitoring: [Data, privacy, and security for Azure OpenAI Service](/legal/cognitive-services/openai/data-privacy?context=/azure/ai-services/openai/context/context#preventing-abuse-and-harmful-content-generation).

articles/ai-services/openai/concepts/model-retirements.md

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ titleSuffix: Azure OpenAI
44
description: Learn about the model deprecations and retirements in Azure OpenAI.
55
ms.service: azure-ai-openai
66
ms.topic: conceptual
7-
ms.date: 09/26/2024
7+
ms.date: 10/02/2024
88
ms.custom:
99
manager: nitinme
1010
author: mrbullwinkle
@@ -91,6 +91,8 @@ These models are currently available for use in Azure OpenAI Service.
9191

9292
| Model | Version | Retirement date | Suggested replacements |
9393
| ---- | ---- | ---- | --- |
94+
| `dall-e-2`| 2 | January 27, 2025 | `dalle-3` |
95+
| `dall-e-3` | 3 | No earlier than April 30, 2025 | |
9496
| `gpt-35-turbo` | 0301 | January 27, 2025<br><br> Deployments set to [**Auto-update to default**](/azure/ai-services/openai/how-to/working-with-models?tabs=powershell#auto-update-to-default) will be automatically upgraded to version: `0125`, starting on November 13, 2024. | `gpt-35-turbo` (0125) <br><br> `gpt-4o-mini` |
9597
| `gpt-35-turbo`<br>`gpt-35-turbo-16k` | 0613 | January 27, 2025 <br><br> Deployments set to [**Auto-update to default**](/azure/ai-services/openai/how-to/working-with-models?tabs=powershell#auto-update-to-default) will be automatically upgraded to version: `0125`, starting on November 13, 2024. | `gpt-35-turbo` (0125) <br><br> `gpt-4o-mini`|
9698
| `gpt-35-turbo` | 1106 | No earlier than January 27, 2025 <br><br> Deployments set to [**Auto-update to default**](/azure/ai-services/openai/how-to/working-with-models?tabs=powershell#auto-update-to-default) will be automatically upgraded to version: `0125`, starting on November 13, 2024. | `gpt-35-turbo` (0125) <br><br> `gpt-4o-mini` |
@@ -100,6 +102,7 @@ These models are currently available for use in Azure OpenAI Service.
100102
| `gpt-4` | 1106-preview | To be upgraded to `gpt-4` version: `turbo-2024-04-09`, starting no sooner than January 27, 2025 **<sup>1</sup>** | `gpt-4o`|
101103
| `gpt-4` | 0125-preview |To be upgraded to `gpt-4` version: `turbo-2024-04-09`, starting no sooner than January 27, 2025 **<sup>1</sup>** | `gpt-4o` |
102104
| `gpt-4` | vision-preview | To be upgraded to `gpt-4` version: `turbo-2024-04-09`, starting no sooner than January 27, 2025 **<sup>1</sup>** | `gpt-4o`|
105+
| `gpt-4o` | 2024-05-13 | No earlier than March 20, 2025 <br><br>Deployments set to [**Auto-update to default**](/azure/ai-services/openai/how-to/working-with-models?tabs=powershell#auto-update-to-default) will be automatically upgraded to version: `2024-08-06`, starting on December 5, 2024. | |
103106
| `gpt-3.5-turbo-instruct` | 0914 | No earlier than Sep 14, 2025 | |
104107
| `text-embedding-ada-002` | 2 | No earlier than April 3, 2025 | `text-embedding-3-small` or `text-embedding-3-large` |
105108
| `text-embedding-ada-002` | 1 | No earlier than April 3, 2025 | `text-embedding-3-small` or `text-embedding-3-large` |

articles/ai-services/openai/concepts/models.md

Lines changed: 20 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ titleSuffix: Azure OpenAI
44
description: Learn about the different model capabilities that are available with Azure OpenAI.
55
ms.service: azure-ai-openai
66
ms.topic: conceptual
7-
ms.date: 09/30/2024
7+
ms.date: 10/01/2024
88
ms.custom: references_regions, build-2023, build-2023-dataai, refefences_regions
99
manager: nitinme
1010
author: mrbullwinkle #ChrisHMSFT
@@ -18,6 +18,7 @@ Azure OpenAI Service is powered by a diverse set of models with different capabi
1818

1919
| Models | Description |
2020
|--|--|
21+
| [o1-preview and o1-mini](#o1-preview-and-o1-mini-models-limited-access) | Limited access models, specifically designed to tackle reasoning and problem-solving tasks with increased focus and capability. |
2122
| [GPT-4o & GPT-4o mini & GPT-4 Turbo](#gpt-4o-and-gpt-4-turbo) | The latest most capable Azure OpenAI models with multimodal versions, which can accept both text and images as input. |
2223
| [GPT-4o audio](#gpt-4o-audio) | A GPT-4o model that supports low-latency, "speech in, speech out" conversational interactions. |
2324
| [GPT-4](#gpt-4) | A set of models that improve on GPT-3.5 and can understand and generate natural language and code. |
@@ -31,18 +32,28 @@ Azure OpenAI Service is powered by a diverse set of models with different capabi
3132

3233
The Azure OpenAI `o1-preview` and `o1-mini` models are specifically designed to tackle reasoning and problem-solving tasks with increased focus and capability. These models spend more time processing and understanding the user's request, making them exceptionally strong in areas like science, coding, and math compared to previous iterations.
3334

34-
### Availability
35+
| Model ID | Description | Max Request (tokens) | Training Data (up to) |
36+
| --- | :--- |:--- |:---: |
37+
|`o1-preview` (2024-09-12) | The most capable model in the o1 series, offering enhanced reasoning abilities.| Input: 128,000 <br> Output: 32,768 | Oct 2023 |
38+
| `o1-mini` (2024-09-12) | A faster and more cost-efficient option in the o1 series, ideal for coding tasks requiring speed and lower resource consumption.| Input: 128,000 <br> Output: 65,536 | Oct 2023 |
3539

36-
The `o1-preview` and `o1-mini` models are available in the East US2 region for limited access through the [AI Studio](https://ai.azure.com) early access playground. Data processing for the `o1` models may occur in a different region than where they are available for use.
40+
### Availability
3741

38-
To try the `o1-preview` and `o1-mini` models in the early access playground, **registration is required, and access will be granted based on Microsofts eligibility criteria**.
42+
The `o1-preview` and `o1-mini` models are now available for API access and model deployment. **Registration is required, and access will be granted based on Microsoft's eligibility criteria**.
3943

4044
Request access: [limited access model application](https://aka.ms/oai/modelaccess)
4145

42-
Once access has been granted, you will need to:
46+
Once access has been granted, you will need to create a deployment for each model.
47+
48+
### API support
49+
50+
Support for the **o1 series** models was added in API version `2024-09-01-preview`.
51+
52+
The `max_tokens` parameter has been deprecated and replaced with the new `max_completion_tokens` parameter. **o1 series** models will only work with the `max_completions_tokens` parameter. `max_completions_tokens` is backwards compatible with `max_tokens`.
53+
54+
### Region availability
4355

44-
1. Navigate to https://ai.azure.com/resources and select a resource in the `eastus2` region. If you do not have an Azure OpenAI resource in this region you will need to [create one](https://portal.azure.com/#create/Microsoft.CognitiveServicesOpenAI).
45-
2. Once the `eastus2` Azure OpenAI resource is selected, in the upper left-hand panel under **Playgrounds** select **Early access playground (preview)**.
56+
Available for standard and global standard deployment in East US2 and Sweden Central for approved customers.
4657

4758
## GPT-4o audio
4859

@@ -289,6 +300,7 @@ The following models support global batch:
289300

290301
| Model | Version | Input format |
291302
|---|---|---|
303+
|`gpt-4o` | 2024-08-06 |text + image |
292304
|`gpt-4o-mini`| 2024-07-18 | text + image |
293305
|`gpt-4o` | 2024-05-13 |text + image |
294306
|`gpt-4` | turbo-2024-04-09 | text |
@@ -421,4 +433,4 @@ For the latest information on model retirements, refer to the [model retirement
421433
- [Model retirement and deprecation](./model-retirements.md)
422434
- [Learn more about working with Azure OpenAI models](../how-to/working-with-models.md)
423435
- [Learn more about Azure OpenAI](../overview.md)
424-
- [Learn more about fine-tuning Azure OpenAI models](../how-to/fine-tuning.md)
436+
- [Learn more about fine-tuning Azure OpenAI models](../how-to/fine-tuning.md)

articles/ai-services/openai/how-to/audio-real-time.md

Lines changed: 24 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
---
2-
title: 'How to use GPT-4o real-time audio with Azure OpenAI Service'
2+
title: 'How to use GPT-4o Realtime API for speech and audio with Azure OpenAI Service'
33
titleSuffix: Azure OpenAI
4-
description: Learn how to use GPT-4o real-time audio with Azure OpenAI Service.
4+
description: Learn how to use GPT-4o Realtime API for speech and audio with Azure OpenAI Service.
55
manager: nitinme
66
ms.service: azure-ai-openai
77
ms.topic: how-to
@@ -12,11 +12,11 @@ ms.custom: references_regions
1212
recommendations: false
1313
---
1414

15-
# GPT-4o real-time audio
15+
# GPT-4o Realtime API for speech and audio (Preview)
1616

17-
Azure OpenAI GPT-4o audio is part of the GPT-4o model family that supports low-latency, "speech in, speech out" conversational interactions. The GPT-4o audio `realtime` API is designed to handle real-time, low-latency conversational interactions, making it a great fit for use cases involving live interactions between a user and a model, such as customer support agents, voice assistants, and real-time translators.
17+
Azure OpenAI GPT-4o Realtime API for speech and audio is part of the GPT-4o model family that supports low-latency, "speech in, speech out" conversational interactions. The GPT-4o audio `realtime` API is designed to handle real-time, low-latency conversational interactions, making it a great fit for use cases involving live interactions between a user and a model, such as customer support agents, voice assistants, and real-time translators.
1818

19-
Most users of this API need to deliver and receive audio from an end-user in real time, including applications that use WebRTC or a telephony system. The real-time API isn't designed to connect directly to end user devices and relies on client integrations to terminate end user audio streams.
19+
Most users of the Realtime API need to deliver and receive audio from an end-user in real time, including applications that use WebRTC or a telephony system. The Realtime API isn't designed to connect directly to end user devices and relies on client integrations to terminate end user audio streams.
2020

2121
## Supported models
2222

@@ -29,7 +29,7 @@ The `gpt-4o-realtime-preview` model is available for global deployments in [East
2929
3030
## API support
3131

32-
Support for real-time audio was first added in API version `2024-10-01-preview`.
32+
Support for the Realtime API was first added in API version `2024-10-01-preview`.
3333

3434
> [!NOTE]
3535
> For more information about the API and architecture, see the [Azure OpenAI GPT-4o real-time audio repository on GitHub](https://github.com/azure-samples/aoai-realtime-audio-sdk).
@@ -43,7 +43,7 @@ Support for real-time audio was first added in API version `2024-10-01-preview`.
4343

4444
Before you can use GPT-4o real-time audio, you need a deployment of the `gpt-4o-realtime-preview` model in a supported region as described in the [supported models](#supported-models) section.
4545

46-
You can deploy the model from the Azure OpenAI model catalog or from your project in AI Studio. Follow these steps to deploy a `gpt-4o-realtime-preview` model from the [AI Studio model catalog](../../../ai-studio/how-to/model-catalog-overview.md):
46+
You can deploy the model from the [Azure AI Studio model catalog](../../../ai-studio/how-to/model-catalog-overview.md) or from your project in AI Studio. Follow these steps to deploy a `gpt-4o-realtime-preview` model from the model catalog:
4747

4848
1. Sign in to [AI Studio](https://ai.azure.com) and go to the **Home** page.
4949
1. Select **Model catalog** from the left sidebar.
@@ -54,17 +54,20 @@ You can deploy the model from the Azure OpenAI model catalog or from your projec
5454
1. Modify other default settings depending on your requirements.
5555
1. Select **Deploy**. You land on the deployment details page.
5656

57-
Now that you have a deployment of the `gpt-4o-realtime-preview` model, you can use the playground to interact with the model in real time. Select **Early access playground** from the list of playgrounds in the left pane.
57+
Now that you have a deployment of the `gpt-4o-realtime-preview` model, you can use the Realtime API to interact with it in real time.
5858

59-
## Use the GPT-4o real-time audio API
59+
## Use the GPT-4o Realtime API
6060

6161
> [!TIP]
6262
> A playground for GPT-4o real-time audio is coming soon to [Azure AI Studio](https://ai.azure.com). You can already use the API directly in your application.
6363
64-
Right now, the fastest way to get started with GPT-4o real-time audio is to download the sample code from the [Azure OpenAI GPT-4o real-time audio repository on GitHub](https://github.com/azure-samples/aoai-realtime-audio-sdk).
64+
Right now, the fastest way to get started with the GPT-4o Realtime API is to download the sample code from the [Azure OpenAI GPT-4o real-time audio repository on GitHub](https://github.com/azure-samples/aoai-realtime-audio-sdk).
65+
66+
The JavaScript web sample demonstrates how to use the GPT-4o Realtime API to interact with the model in real time. The sample code includes a simple web interface that captures audio from the user's microphone and sends it to the model for processing. The model responds with text and audio, which the sample code renders in the web interface.
67+
68+
You can run the sample code locally on your machine by following these steps. Refer to the [repository on GitHub](https://github.com/azure-samples/aoai-realtime-audio-sdk) for the most up-to-date instructions.
69+
1. If you don't have Node.js installed, download and install the [LTS version of Node.js](https://nodejs.org/).
6570

66-
The JavaScript web sample demonstrates how to use the GPT-4o real-time audio API to interact with the model in real time. The sample code includes a simple web interface that captures audio from the user's microphone and sends it to the model for processing. The model responds with text and audio, which the sample code renders in the web interface.
67-
6871
1. Clone the repository to your local machine:
6972

7073
```bash
@@ -74,12 +77,18 @@ The JavaScript web sample demonstrates how to use the GPT-4o real-time audio API
7477
1. Go to the `javascript/samples/web` folder in your preferred code editor.
7578

7679
```bash
77-
cd .\javascript\samples\web\
80+
cd ./javascript/samples
7881
```
7982

80-
1. If you don't have Node.js installed, download and install the [LTS version of Node.js](https://nodejs.org/).
83+
1. Run `download-pkg.ps1` or `download-pkg.sh` to download the required packages.
84+
85+
1. Go to the `web` folder from the `./javascript/samples` folder.
86+
87+
```bash
88+
cd ./web
89+
```
8190

82-
1. Run `npm install` to download a few dependency packages. For more information, see the `package.json` file in the same `web` folder.
91+
1. Run `npm install` to install package dependencies.
8392

8493
1. Run `npm run dev` to start the web server, navigating any firewall permissions prompts as needed.
8594
1. Go to any of the provided URIs from the console output (such as `http://localhost:5173/`) in a browser.

0 commit comments

Comments
 (0)