You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@@ -26,24 +26,6 @@ Azure OpenAI Service is powered by a diverse set of models with different capabi
26
26
|[Whisper](#whisper-models)| A series of models in preview that can transcribe and translate speech to text. |
27
27
|[Text to speech](#text-to-speech-models-preview) (Preview) | A series of models in preview that can synthesize text to speech. |
28
28
29
-
## Early access playground (preview)
30
-
31
-
On August 6, 2024, OpenAI [announced](https://openai.com/index/introducing-structured-outputs-in-the-api/) the latest version of their flagship GPT-4o model version `2024-08-06`. GPT-4o `2024-08-06` has all the capabilities of the previous version as well as:
32
-
33
-
* An enhanced ability to support complex structured outputs.
34
-
* Max output tokens have been increased from 4,096 to 16,384.
35
-
36
-
Azure customers can test out GPT-4o `2024-08-06` today in the new AI Studio early access playground (preview).
37
-
38
-
Unlike the previous early access playground, the AI Studio early access playground (preview) does not require you to have a resource in a specific region.
39
-
40
-
> [!NOTE]
41
-
> Prompts and completions made through the early access playground (preview) may be processed in any Azure OpenAI region, and are currently subject to a 10 request per minute per Azure subscription limit. This limit may change in the future.
42
-
>
43
-
> Azure OpenAI Service abuse monitoring is enabled for all early access playground users even if approved for modification; default content filters are enabled and cannot be modified.
44
-
45
-
To test out GPT-4o `2024-08-06`, sign-in to the Azure AI early access playground (preview) using this [link](https://aka.ms/oai/docs/earlyaccessplayground).
46
-
47
29
## GPT-4o and GPT-4 Turbo
48
30
49
31
GPT-4o integrates text and images in a single model, enabling it to handle multiple data types simultaneously. This multimodal approach enhances accuracy and responsiveness in human-computer interactions. GPT-4o matches GPT-4 Turbo in English text and coding tasks while offering superior performance in non-English languages and vision tasks, setting new benchmarks for AI capabilities.
@@ -56,6 +38,7 @@ You need to [create](../how-to/create-resource.md) or use an existing resource i
56
38
57
39
When your resource is created, you can [deploy](../how-to/create-resource.md#deploy-a-model) the GPT-4o models. If you are performing a programmatic deployment, the **model** names are:
58
40
41
+
-`gpt-4o`**Version**`2024-08-06`
59
42
-`gpt-4o`, **Version**`2024-05-13`
60
43
-`gpt-4o-mini`**Version**`2024-07-18`
61
44
@@ -83,8 +66,9 @@ See [model versions](../concepts/model-versions.md) to learn about how Azure Ope
83
66
84
67
| Model ID | Description | Max Request (tokens) | Training Data (up to) |
85
68
| --- | :--- |:--- |:---: |
69
+
|`gpt-4o` (2024-08-06) <br> **GPT-4o (Omni)**|**Latest large GA model** <br> - Structured outputs<br> - Text, image processing <br> - JSON Mode <br> - parallel function calling <br> - Enhanced accuracy and responsiveness <br> - Parity with English text and coding tasks compared to GPT-4 Turbo with Vision <br> - Superior performance in non-English languages and in vision tasks |Input: 128,000 <br> Output: 4,096| Oct 2023 |
86
70
|`gpt-4o-mini` (2024-07-18) <br> **GPT-4o mini**|**Latest small GA model** <br> - Fast, inexpensive, capable model ideal for replacing GPT-3.5 Turbo series models. <br> - Text, image processing <br>- JSON Mode <br> - parallel function calling | Input: 128,000 <br> Output: 16,384 | Oct 2023 |
87
-
|`gpt-4o` (2024-05-13) <br> **GPT-4o (Omni)**|**Latest large GA model** <br> - Text, image processing <br> - JSON Mode <br> - parallel function calling <br> - Enhanced accuracy and responsiveness <br> - Parity with English text and coding tasks compared to GPT-4 Turbo with Vision <br> - Superior performance in non-English languages and in vision tasks |Input: 128,000 <br> Output: 4,096| Oct 2023 |
71
+
|`gpt-4o` (2024-05-13) <br> **GPT-4o (Omni)**| Text, image processing <br> - JSON Mode <br> - parallel function calling <br> - Enhanced accuracy and responsiveness <br> - Parity with English text and coding tasks compared to GPT-4 Turbo with Vision <br> - Superior performance in non-English languages and in vision tasks |Input: 128,000 <br> Output: 4,096| Oct 2023 |
88
72
|`gpt-4` (turbo-2024-04-09) <br>**GPT-4 Turbo with Vision**|**New GA model** <br> - Replacement for all previous GPT-4 preview models (`vision-preview`, `1106-Preview`, `0125-Preview`). <br> - [**Feature availability**](#gpt-4o-and-gpt-4-turbo) is currently different depending on method of input, and deployment type. | Input: 128,000 <br> Output: 4,096 | Dec 2023 |
89
73
|`gpt-4` (0125-Preview)*<br>**GPT-4 Turbo Preview**|**Preview Model** <br> -Replaces 1106-Preview <br>- Better code generation performance <br> - Reduces cases where the model doesn't complete a task <br> - JSON Mode <br> - parallel function calling <br> - reproducible output (preview) | Input: 128,000 <br> Output: 4,096 | Dec 2023 |
Copy file name to clipboardExpand all lines: articles/ai-services/openai/how-to/json-mode.md
+7-2Lines changed: 7 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -6,7 +6,7 @@ services: cognitive-services
6
6
manager: nitinme
7
7
ms.service: azure-ai-openai
8
8
ms.topic: how-to
9
-
ms.date: 12/05/2023
9
+
ms.date: 08/29/2024
10
10
author: mrbullwinkle
11
11
ms.author: mbullwin
12
12
recommendations: false
@@ -17,6 +17,9 @@ recommendations: false
17
17
18
18
JSON mode allows you to set the models response format to return a valid JSON object as part of a chat completion. While generating valid JSON was possible previously, there could be issues with response consistency that would lead to invalid JSON objects being generated.
19
19
20
+
> [!NOTE]
21
+
> While JSON mode is still supported, when possible we recommend using [structured outputs](./structured-outputs.md). Like JSON mode structured outputs generates valid JSON, but with the added benefit that you can constrain the model to use a specific JSON schema.
22
+
20
23
## JSON mode support
21
24
22
25
JSON mode is only currently supported with the following models:
@@ -27,6 +30,8 @@ JSON mode is only currently supported with the following models:
27
30
*`gpt-35-turbo` (0125)
28
31
*`gpt-4` (1106-Preview)
29
32
*`gpt-4` (0125-Preview)
33
+
*`gpt-4o`
34
+
*`gpt-4o-mini`
30
35
31
36
### API support
32
37
@@ -47,7 +52,7 @@ client = AzureOpenAI(
47
52
)
48
53
49
54
response = client.chat.completions.create(
50
-
model="gpt-4-0125-Preview", # Model = should match the deployment name you chose for your 0125-Preview model deployment
55
+
model="YOUR-MODEL_DEPLOYMENT_NAME", # Model = should match the deployment name you chose for your model deployment
51
56
response_format={ "type": "json_object" },
52
57
messages=[
53
58
{"role": "system", "content": "You are a helpful assistant designed to output JSON."},
0 commit comments