You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-services/openai/whats-new.md
+19-19Lines changed: 19 additions & 19 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -26,7 +26,7 @@ The `o1-preview` and `o1-mini` models are now available for API access and model
26
26
27
27
Request access: [limited access model application](https://aka.ms/oai/modelaccess)
28
28
29
-
Customers who were already approved and have access to the model through the early access playground do not need to apply again, you will automatically be granted API access. Once access has been granted, you will need to create a deployment for each model.
29
+
Customers who were already approved and have access to the model through the early access playground don't need to apply again, you'll automatically be granted API access. Once access has been granted, you'll need to create a deployment for each model.
30
30
31
31
**API support:**
32
32
@@ -58,7 +58,7 @@ On September 19, when you access the [Azure OpenAI Studio](https://oai.azure.com
58
58
59
59
60
60
### GPT-4o 2024-08-06 provisioned deployments
61
-
GPT-4o 2024-08-06 is now available for provisioned deployments in East US, East US 2, North Central US, and Sweden Central. It is also available for global provisioned deployments.
61
+
GPT-4o 2024-08-06 is now available for provisioned deployments in East US, East US 2, North Central US, and Sweden Central. It's also available for global provisioned deployments.
62
62
63
63
For the latest information on model availability, see the [models page](/azure/ai-services/openai/concepts/models#provisioned-deployment-model-availability).
64
64
@@ -69,7 +69,7 @@ For more information, see the [deployment types guide](https://aka.ms/aoai/docs/
69
69
70
70
### NEW o1-preview and o1-mini models available for limited access
71
71
72
-
The Azure OpenAI `o1-preview` and `o1-mini` models are specifically designed to tackle reasoning and problem-solving tasks with increased focus and capability. These models spend more time processing and understanding the user's request, making them exceptionally strong in areas like science, coding, and math compared to previous iterations.
72
+
The Azure OpenAI `o1-preview` and `o1-mini` models are designed to tackle reasoning and problem-solving tasks with increased focus and capability. These models spend more time processing and understanding the user's request, making them exceptionally strong in areas like science, coding, and math compared to previous iterations.
73
73
74
74
### Key capabilities of the o1 series
75
75
@@ -89,23 +89,23 @@ Request access: [limited access model application](https://aka.ms/oai/modelacces
89
89
90
90
### Limitations
91
91
92
-
The `o1` series models are currently in preview and do not include some features available in other models, such as image understanding and structured outputs which are available in the latest GPT-4o model. For many tasks, the generally available GPT-4o models may still be more suitable.
92
+
The `o1` series models are currently in preview and don't include some features available in other models, such as image understanding and structured outputs which are available in the latest GPT-4o model. For many tasks, the generally available GPT-4o models might still be more suitable.
93
93
94
94
### Safety
95
95
96
96
OpenAI has incorporated additional safety measures into the `o1` models, including new techniques to help the models refuse unsafe requests. These advancements make the `o1` series some of the most robust models available.
97
97
98
98
### Availability
99
99
100
-
The `o1-preview` and `o1-mini` are available in the East US2 region for limited access through the [AI Studio](https://ai.azure.com) early access playground. Data processing for the `o1` models may occur in a different region than where they are available for use.
100
+
The `o1-preview` and `o1-mini` are available in the East US2 region for limited access through the [AI Studio](https://ai.azure.com) early access playground. Data processing for the `o1` models might occur in a different region than where they are available for use.
101
101
102
102
To try the `o1-preview` and `o1-mini` models in the early access playground **registration is required, and access will be granted based on Microsoft’s eligibility criteria.**
103
103
104
104
Request access: [limited access model application](https://aka.ms/oai/modelaccess)
105
105
106
106
Once access has been granted, you will need to:
107
107
108
-
1. Navigate to https://ai.azure.com/resources and select a resource in the `eastus2` region. If you do not have an Azure OpenAI resource in this region you will need to [create one](https://portal.azure.com/#create/Microsoft.CognitiveServicesOpenAI).
108
+
1. Navigate to https://ai.azure.com/resources and select a resource in the `eastus2` region. If you don't have an Azure OpenAI resource in this region you'll need to [create one](https://portal.azure.com/#create/Microsoft.CognitiveServicesOpenAI).
109
109
2. Once the `eastus2` Azure OpenAI resource is selected, in the upper left-hand panel under **Playgrounds** select **Early access playground (preview)**.
110
110
111
111
## August 2024
@@ -155,10 +155,10 @@ On August 6, 2024, OpenAI [announced](https://openai.com/index/introducing-struc
155
155
156
156
Azure customers can test out GPT-4o `2024-08-06` today in the new AI Studio early access playground (preview).
157
157
158
-
Unlike the previous early access playground, the AI Studio early access playground (preview) does not require you to have a resource in a specific region.
158
+
Unlike the previous early access playground, the AI Studio early access playground (preview) doesn't require you to have a resource in a specific region.
159
159
160
160
> [!NOTE]
161
-
> Prompts and completions made through the early access playground (preview) may be processed in any Azure OpenAI region, and are currently subject to a 10 request per minute per Azure subscription limit. This limit may change in the future.
161
+
> Prompts and completions made through the early access playground (preview) might be processed in any Azure OpenAI region, and are currently subject to a 10 request per minute per Azure subscription limit. This limit might change in the future.
162
162
>
163
163
> Azure OpenAI Service abuse monitoring is enabled for all early access playground users even if approved for modification; default content filters are enabled and cannot be modified.
164
164
@@ -287,7 +287,7 @@ Threads and Files in Assistants now supports CMK in the following region:
287
287
### GPT-4o provisioned deployments
288
288
289
289
`gpt-4o` Version: `2024-05-13` is available for both standard and provisioned deployments. Provisioned and standard model deployments accept both text and image/vision inference requests.
290
-
For information on model regional availability consult the model matrix for [provisioned deployments](./concepts/models.md#provisioned-deployment-model-availability).
290
+
For information on model regional availability, consult the model matrix for [provisioned deployments](./concepts/models.md#provisioned-deployment-model-availability).
291
291
292
292
### Assistants v2 (preview)
293
293
@@ -324,7 +324,7 @@ For information on model regional availability, see the [models page](./concepts
324
324
325
325
### Global standard deployment type (preview)
326
326
327
-
Global deployments are available in the same Azure OpenAI resources as non-global offers but allow you to leverage Azure's global infrastructure to dynamically route traffic to the data center with best availability for each request. Global standard will provide the highest default quota for new models and eliminates the need to load balance across multiple resources.
327
+
Global deployments are available in the same Azure OpenAI resources as non-global offers but allow you to leverage Azure's global infrastructure to dynamically route traffic to the data center with best availability for each request. Global standard provides the highest default quota for new models and eliminates the need to load balance across multiple resources.
328
328
329
329
For more information, see the [deployment types guide](https://aka.ms/aoai/docs/deployment-types).
330
330
@@ -343,7 +343,7 @@ Running filters asynchronously for improved latency in streaming scenarios is no
343
343
344
344
### Prompt Shields
345
345
346
-
Prompt Shields protect applications powered by Azure OpenAI models from two types of attacks: direct (jailbreak) and indirect attacks. Indirect Attacks (also known as Indirect Prompt Attacks or Cross-Domain Prompt Injection Attacks) are a type of attack on systems powered by Generative AI models that may occur when an application processes information that wasn’t directly authored by either the developer of the application or the user. [Content filtering](/azure/ai-services/openai/concepts/content-filter?tabs=warning%2Cpython-new#prompt-shields)
346
+
Prompt Shields protect applications powered by Azure OpenAI models from two types of attacks: direct (jailbreak) and indirect attacks. Indirect Attacks (also known as Indirect Prompt Attacks or Cross-Domain Prompt Injection Attacks) are a type of attack on systems powered by Generative AI models that might occur when an application processes information that wasn’t directly authored by either the developer of the application or the user. [Content filtering](/azure/ai-services/openai/concepts/content-filter?tabs=warning%2Cpython-new#prompt-shields)
347
347
348
348
### 2024-05-01-preview API release
349
349
@@ -400,7 +400,7 @@ Azure OpenAI Studio now provides a Risks & Safety dashboard for each of your dep
400
400
401
401
This is the latest GA API release and is the replacement for the previous `2023-05-15` GA release. This release adds support for the latest Azure OpenAI GA features like Whisper, DALLE-3, fine-tuning, on your data, etc.
402
402
403
-
Features that are still in preview such as Assistants, text to speech (TTS), certain on your data datasources, still require a preview API version. For more information check out our [API version lifecycle guide](./api-version-deprecation.md).
403
+
Features that are still in preview such as Assistants, text to speech (TTS), certain on your data datasources, still require a preview API version. For more information, check out our [API version lifecycle guide](./api-version-deprecation.md).
404
404
405
405
### Whisper general availability (GA)
406
406
@@ -427,9 +427,9 @@ We have added a page to track [model deprecations and retirements](./concepts/mo
427
427
`2024-03-01-preview` has all the same functionality as `2024-02-15-preview` and adds two new parameters for embeddings:
428
428
429
429
-`encoding_format` allows you to specify the format to generate embeddings in `float`, or `base64`. The default is `float`.
430
-
-`dimensions` allows you set the number of output embeddings. This parameter is only supported with the new third generation embeddings models: `text-embedding-3-large`, `text-embedding-3-small`. Typically larger embeddings are more expensive from a compute, memory, and storage perspective. Being able to adjust the number of dimensions allows more control over overall cost and performance. The `dimensions` parameter is not supported in all versions of the OpenAI 1.x Python library, to take advantage of this parameter we recommend upgrading to the latest version: `pip install openai --upgrade`.
430
+
-`dimensions` allows you set the number of output embeddings. This parameter is only supported with the new third generation embeddings models: `text-embedding-3-large`, `text-embedding-3-small`. Typically larger embeddings are more expensive from a compute, memory, and storage perspective. Being able to adjust the number of dimensions allows more control over overall cost and performance. The `dimensions` parameter isn't supported in all versions of the OpenAI 1.x Python library, to take advantage of this parameter we recommend upgrading to the latest version: `pip install openai --upgrade`.
431
431
432
-
If you are currently using a preview API version to take advantage of the latest features, we recommend consulting the [API version lifecycle](./api-version-deprecation.md) article to track how long your current API version will be supported.
432
+
If you're currently using a preview API version to take advantage of the latest features, we recommend consulting the [API version lifecycle](./api-version-deprecation.md) article to track how long your current API version will be supported.
433
433
434
434
### Update to GPT-4-1106-Preview upgrade plans
435
435
@@ -456,7 +456,7 @@ For information on model regional availability and upgrades refer to the [models
456
456
457
457
### GPT-3.5 Turbo quota consolidation
458
458
459
-
To simplify migration between different versions of the GPT-3.5-Turbo models (including 16k), we will be consolidating all GPT-3.5-Turbo quota into a single quota value.
459
+
To simplify migration between different versions of the GPT-3.5-Turbo models (including 16k), we'll be consolidating all GPT-3.5-Turbo quota into a single quota value.
460
460
461
461
- Any customers who have increased quota approved will have combined total quota that reflects the previous increases.
462
462
@@ -524,13 +524,13 @@ GPT-4 Turbo with Vision on Azure OpenAI service is now in public preview. GPT-4
524
524
525
525
### New data source support in Azure OpenAI On Your Data
526
526
527
-
- You can now use [Azure Cosmos DB for MongoDB vCore](./concepts/use-your-data.md#supported-data-sources)as well as URLs/web addresses as data sources to ingest your data and chat with a supported Azure OpenAI model.
527
+
- You can now use [Azure Cosmos DB for MongoDB vCore](./concepts/use-your-data.md#supported-data-sources)and URLs/web addresses as data sources to ingest your data and chat with a supported Azure OpenAI model.
528
528
529
529
### GPT-4 Turbo Preview & GPT-3.5-Turbo-1106 released
530
530
531
531
Both models are the latest release from OpenAI with improved instruction following, [JSON mode](./how-to/json-mode.md), [reproducible output](./how-to/reproducible-output.md), and parallel function calling.
532
532
533
-
-**GPT-4 Turbo Preview** has a max context window of 128,000 tokens and can generate 4,096 output tokens. It has the latest training data with knowledge up to April 2023. This model is in preview and is not recommended for production use. All deployments of this preview model will be automatically updated in place once the stable release becomes available.
533
+
-**GPT-4 Turbo Preview** has a max context window of 128,000 tokens and can generate 4,096 output tokens. It has the latest training data with knowledge up to April 2023. This model is in preview and isn't recommended for production use. All deployments of this preview model will be automatically updated in place once the stable release becomes available.
534
534
535
535
-**GPT-3.5-Turbo-1106** has a max context window of 16,385 tokens and can generate 4,096 output tokens.
536
536
@@ -671,7 +671,7 @@ Azure OpenAI Service now supports speech to text APIs powered by OpenAI's Whispe
671
671
- GPT-35-Turbo models.
672
672
- GPT-4 model series.
673
673
674
-
If you are currently using the `2023-03-15-preview` API, we recommend migrating to the GA `2023-05-15` API. If you are currently using API version `2022-12-01` this API remains GA, but does not include the latest Chat Completion capabilities.
674
+
If you're currently using the `2023-03-15-preview` API, we recommend migrating to the GA `2023-05-15` API. If you're currently using API version `2022-12-01` this API remains GA, but doesn't include the latest Chat Completion capabilities.
675
675
676
676
> [!IMPORTANT]
677
677
> Using the current versions of the GPT-35-Turbo models with the completion endpoint remains in preview.
@@ -696,7 +696,7 @@ If you are currently using the `2023-03-15-preview` API, we recommend migrating
696
696
-**GPT-35-Turbo preview**. To learn more checkout the [how-to article](./how-to/chatgpt.md).
697
697
698
698
- Increased training limits for fine-tuning: The max training job size (tokens in training file) x (# of epochs) is 2 Billion tokens for all models. We have also increased the max training job from 120 to 720 hours.
699
-
- Adding additional use cases to your existing access. Previously, the process for adding new use cases required customers to reapply to the service. Now, we're releasing a new process that allows you to quickly add new use cases to your use of the service. This process follows the established Limited Access process within Azure AI services. [Existing customers can attest to any and all new use cases here](https://customervoice.microsoft.com/Pages/ResponsePage.aspx?id=v4j5cvGGr0GRqy180BHbR7en2Ais5pxKtso_Pz4b1_xUM003VEJPRjRSOTZBRVZBV1E5N1lWMk1XUyQlQCN0PWcu). Please note that this is required anytime you would like to use the service for a new use case you did not originally apply for.
699
+
- Adding additional use cases to your existing access. Previously, the process for adding new use cases required customers to reapply to the service. Now, we're releasing a new process that allows you to quickly add new use cases to your use of the service. This process follows the established Limited Access process within Azure AI services. [Existing customers can attest to any and all new use cases here](https://customervoice.microsoft.com/Pages/ResponsePage.aspx?id=v4j5cvGGr0GRqy180BHbR7en2Ais5pxKtso_Pz4b1_xUM003VEJPRjRSOTZBRVZBV1E5N1lWMk1XUyQlQCN0PWcu). Please note that this is required anytime you would like to use the service for a new use case you didn't originally apply for.
0 commit comments