You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@@ -26,6 +26,23 @@ Azure OpenAI Service is powered by a diverse set of models with different capabi
26
26
|[Whisper](#whisper-models)| A series of models in preview that can transcribe and translate speech to text. |
27
27
|[Text to speech](#text-to-speech-models-preview) (Preview) | A series of models in preview that can synthesize text to speech. |
28
28
29
+
## o1-preview and o1-mini models limited access
30
+
31
+
The Azure OpenAI `o1-preview` and `o1-mini` models are specifically designed to tackle reasoning and problem-solving tasks with increased focus and capability. These models spend more time processing and understanding the user's request, making them exceptionally strong in areas like science, coding, and math compared to previous iterations.
32
+
33
+
### Availability
34
+
35
+
The `o1-preview` and `o1-mini` models are available in the East US2 region for limited access through the [AI Studio](https://ai.azure.com) early access playground. Data processing for the `o1` models may occur in a different region than where they are available for use.
36
+
37
+
To try the `o1-preview` and `o1-mini` models in the early access playground, **registration is required, and access will be granted based on Microsoft’s eligibility criteria**.
38
+
39
+
Request access: [limited access model application](https://aka.ms/oai/modelaccess)
40
+
41
+
Once access has been granted, you will need to:
42
+
43
+
1. Navigate to https://ai.azure.com/resources and select a resource in the `eastus2` region. If you do not have an Azure OpenAI resource in this region you will need to [create one](https://portal.azure.com/#create/Microsoft.CognitiveServicesOpenAI).
44
+
2. Once the `eastus2` Azure OpenAI resource is selected, in the upper left-hand panel under **Playgrounds** select **Early access playground (preview)**.
45
+
29
46
## GPT-4o and GPT-4 Turbo
30
47
31
48
GPT-4o integrates text and images in a single model, enabling it to handle multiple data types simultaneously. This multimodal approach enhances accuracy and responsiveness in human-computer interactions. GPT-4o matches GPT-4 Turbo in English text and coding tasks while offering superior performance in non-English languages and vision tasks, setting new benchmarks for AI capabilities.
Copy file name to clipboardExpand all lines: articles/ai-services/openai/whats-new.md
+44-1Lines changed: 44 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -10,14 +10,57 @@ ms.custom:
10
10
- ignite-2023
11
11
- references_regions
12
12
ms.topic: whats-new
13
-
ms.date: 9/03/2024
13
+
ms.date: 9/12/2024
14
14
recommendations: false
15
15
---
16
16
17
17
# What's new in Azure OpenAI Service
18
18
19
19
This article provides a summary of the latest releases and major documentation updates for Azure OpenAI.
20
20
21
+
## September 2024
22
+
23
+
### NEW o1-preview and o1-mini models available for limited access
24
+
25
+
The Azure OpenAI `o1-preview` and `o1-mini` models are specifically designed to tackle reasoning and problem-solving tasks with increased focus and capability. These models spend more time processing and understanding the user's request, making them exceptionally strong in areas like science, coding, and math compared to previous iterations.
26
+
27
+
### Key capabilities of the o1 series
28
+
29
+
- Complex Code Generation: Capable of generating algorithms and handling advanced coding tasks to support developers.
30
+
- Advanced Problem Solving: Ideal for comprehensive brainstorming sessions and addressing multifaceted challenges.
31
+
- Complex Document Comparison: Perfect for analyzing contracts, case files, or legal documents to identify subtle differences.
32
+
- Instruction Following and Workflow Management: Particularly effective for managing workflows requiring shorter contexts.
33
+
34
+
### Model variants
35
+
36
+
-`o1-preview`: `o1-preview` is the more capable of the `o1` series models.
37
+
-`o1-mini`: `o1-mini` is the faster and cheaper of the `o1` series models.
38
+
39
+
Model version: `2024-09-12`
40
+
41
+
Request access: [limited access model application](https://aka.ms/oai/modelaccess)
42
+
43
+
### Limitations
44
+
45
+
The `o1-preview` and `o1-mini` models are currently in preview and do not include some features available in other models, such as image understanding and structured outputs found in the GPT-4o and GPT-4o-mini models. For many tasks, the generally available GPT-4o models may still be more suitable.
46
+
47
+
### Safety
48
+
49
+
OpenAI has incorporated additional safety measures into the `o1` models, including new techniques to help the models refuse unsafe requests. These advancements make the `o1` series some of the most robust models available.
50
+
51
+
### Availability
52
+
53
+
The `o1-preview` and `o1-mini` are available in the East US2 region for limited access through the [AI Studio](https://ai.azure.com) early access playground. Data processing for the `o1` models may occur in a different region than where they are available for use.
54
+
55
+
To try the `o1-preview` and `o1-mini` models in the early access playground **registration is required, and access will be granted based on Microsoft’s eligibility criteria.**
56
+
57
+
Request access: [limited access model application](https://aka.ms/oai/modelaccess)
58
+
59
+
Once access has been granted, you will need to:
60
+
61
+
1. Navigate to https://ai.azure.com/resources and select a resource in the `eastus2` region. If you do not have an Azure OpenAI resource in this region you will need to [create one](https://portal.azure.com/#create/Microsoft.CognitiveServicesOpenAI).
62
+
2. Once the `eastus2` Azure OpenAI resource is selected, in the upper left-hand panel under **Playgrounds** select **Early access playground (preview)**.
Copy file name to clipboardExpand all lines: articles/ai-studio/concepts/rbac-ai-studio.md
+11-11Lines changed: 11 additions & 11 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -8,7 +8,7 @@ ms.custom:
8
8
- ignite-2023
9
9
- build-2024
10
10
ms.topic: conceptual
11
-
ms.date: 5/21/2024
11
+
ms.date: 9/12/2024
12
12
ms.reviewer: deeikele
13
13
ms.author: larryfr
14
14
author: Blackmist
@@ -220,20 +220,20 @@ When you create a connection that uses Microsoft Entra ID authentication, you mu
220
220
221
221
| Resource connection | Role | Description |
222
222
|----------|------|-------------|
223
-
| Azure AI Search | Contributor | List API-Keys to list indexes from Azure OpenAI Studio. |
223
+
| Azure AI Search | Contributor | List API-Keys to list indexes from Azure AI Studio. |
224
224
| Azure AI Search | Search Index Data Contributor | Required for indexing scenarios |
225
-
| Azure AI services/Azure OpenAI | Cognitive Services OpenAI Contributor | Call public ingestion API from Azure OpenAI Studio. |
226
-
| Azure AI services/OpenAI | Cognitive Services User | List API-Keys from Azure OpenAI Studio. |
227
-
| Azure AI services/OpenAI | Contributor | Allows for calls to the control plane. |
225
+
| Azure AI services / Azure OpenAI | Cognitive Services OpenAI Contributor | Call public ingestion API from Azure AI Studio. |
226
+
| Azure AI services / Azure OpenAI | Cognitive Services User | List API-Keys from Azure AI Studio. |
227
+
| Azure AI services / Azure OpenAI | Contributor | Allows for calls to the control plane. |
228
228
229
229
When using Microsoft Entra ID authenticated connections in the chat playground, the services need to authorize each other to access the required resources. The admin performing the configuration needs to have the __Owner__ role on these resources to add role assignments. The following table lists the required role assignments for each resource. The __Assignee__ column refers to the system-assigned managed identity of the listed resource. The __Resource__ column refers to the resource that the assignee needs to access. For example, Azure OpenAI has a system-assigned managed identity that needs to be assigned the __Search Index Data Reader__ role for the Azure AI Search resource.
230
230
231
231
| Role | Assignee | Resource | Description |
232
232
|------|----------|----------|-------------|
233
-
| Search Index Data Reader | Azure AI services/OpenAI | Azure AI Search | Inference service queries the data from the index. Only used for inference scenarios. |
234
-
| Search Index Data Contributor | Azure AI services/OpenAI | Azure AI Search | Read-write access to content in indexes. Import, refresh, or query the documents collection of an index. Only used for ingestion and inference scenarios. |
235
-
| Search Service Contributor | Azure AI services/OpenAI | Azure AI Search | Read-write access to object definitions (indexes, aliases, synonym maps, indexers, data sources, and skillsets). Inference service queries the index schema for auto fields mapping. Data ingestion service creates index, data sources, skill set, indexer, and queries the indexer status. |
236
-
| Cognitive Services OpenAI Contributor | Azure AI Search | Azure AI services/OpenAI | Custom skill |
233
+
| Search Index Data Reader | Azure AI services / Azure OpenAI | Azure AI Search | Inference service queries the data from the index. Only used for inference scenarios. |
234
+
| Search Index Data Contributor | Azure AI services / Azure OpenAI | Azure AI Search | Read-write access to content in indexes. Import, refresh, or query the documents collection of an index. Only used for ingestion and inference scenarios. |
235
+
| Search Service Contributor | Azure AI services / Azure OpenAI | Azure AI Search | Read-write access to object definitions (indexes, aliases, synonym maps, indexers, data sources, and skillsets). Inference service queries the index schema for auto fields mapping. Data ingestion service creates index, data sources, skill set, indexer, and queries the indexer status. |
236
+
| Cognitive Services OpenAI Contributor | Azure AI Search | Azure AI services / Azure OpenAI | Custom skill |
237
237
| Cognitive Services OpenAI User | Azure OpenAI Resource for chat model | Azure OpenAI resource for embedding model | Required only if using two Azure OpenAI resources to communicate. |
238
238
239
239
> [!NOTE]
@@ -316,8 +316,8 @@ The following example defines a role for a developer using [Azure OpenAI Assista
316
316
{
317
317
"id": "",
318
318
"properties": {
319
-
"roleName": "CognitiveServices OpenAI Assistants API Developer",
320
-
"description": "Custom role to work with AOAI Assistants API",
319
+
"roleName": "Azure OpenAI Assistants API Developer",
320
+
"description": "Custom role to work with Azure OpenAI Assistants API",
Copy file name to clipboardExpand all lines: articles/ai-studio/how-to/deploy-models-jais.md
+11-11Lines changed: 11 additions & 11 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -201,7 +201,7 @@ print_stream(result)
201
201
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](https://aka.ms/azureai/modelinference).
202
202
203
203
```python
204
-
from azure.ai.inference.models importChatCompletionsResponseFormatText
204
+
from azure.ai.inference.models importChatCompletionsResponseFormat
> Jais doesn't support JSON output formatting (`response_format = { "type": "json_object" }`). You can always prompt the model to generate JSON outputs. However, such outputs are not guaranteed to be valid JSON.
222
+
> Jais models don't support JSON output formatting (`response_format = { "type": "json_object" }`). You can always prompt the model to generate JSON outputs. However, such outputs are not guaranteed to be valid JSON.
223
223
224
224
If you want to pass a parameter that isn't in the list of supported parameters, you can pass it to the underlying model using *extra parameters*. See [Pass extra parameters to the model](#pass-extra-parameters-to-the-model).
225
225
@@ -482,7 +482,7 @@ var response = await client.path("/chat/completions").post({
482
482
```
483
483
484
484
> [!WARNING]
485
-
> Jais doesn't support JSON output formatting (`response_format = { "type":"json_object" }`). You can always prompt the model to generate JSON outputs. However, such outputs are not guaranteed to be valid JSON.
485
+
> Jais models don't support JSON output formatting (`response_format = { "type":"json_object" }`). You can always prompt the model to generate JSON outputs. However, such outputs are not guaranteed to be valid JSON.
486
486
487
487
If you want to pass a parameter that isn't in the list of supported parameters, you can pass it to the underlying model using *extra parameters*. See [Pass extra parameters to the model](#pass-extra-parameters-to-the-model).
488
488
@@ -580,7 +580,7 @@ Deployment to a serverless API endpoint doesn't require quota from your subscrip
580
580
581
581
### The inference package installed
582
582
583
-
You can consume predictions from this model by using the `Azure.AI.Inference` package from [Nuget](https://www.nuget.org/). To install this package, you need the following prerequisites:
583
+
You can consume predictions from this model by using the `Azure.AI.Inference` package from [NuGet](https://www.nuget.org/). To install this package, you need the following prerequisites:
584
584
585
585
* The endpoint URL. To construct the client library, you need to pass in the endpoint URL. The endpoint URL has the form `https://your-host-name.your-azure-region.inference.ai.azure.com`, where `your-host-name` is your unique model deployment host name and `your-azure-region` is the Azure region where the model is deployed (for example, eastus2).
586
586
* Depending on your model deployment and authentication preference, you need either a key to authenticate against the service, or Microsoft Entra IDcredentials. The key is a 32-character string.
@@ -606,7 +606,7 @@ using Azure.Identity;
606
606
using Azure.AI.Inference;
607
607
```
608
608
609
-
This example also use the following namespaces but you may not always need them:
609
+
This example also uses the following namespaces but you may not always need them:
> Jais doesn't support JSON output formatting (`response_format = { "type": "json_object" }`). You can always prompt the model to generate JSONoutputs. However, such outputs are not guaranteed to be valid JSON.
778
+
> Jais models don't support JSON output formatting (`response_format = { "type": "json_object" }`). You can always prompt the model to generate JSONoutputs. However, such outputs are not guaranteed to be valid JSON.
779
779
780
780
If you want to pass a parameter that isn't in the list of supported parameters, you can pass it to the underlying model using *extra parameters*. See [Pass extra parameters to the model](#pass-extra-parameters-to-the-model).
781
781
@@ -1088,7 +1088,7 @@ Explore other parameters that you can specify in the inference client. For a ful
1088
1088
```
1089
1089
1090
1090
> [!WARNING]
1091
-
> Jais doesn't support JSON output formatting (`response_format = { "type": "json_object" }`). You can always prompt the model to generate JSON outputs. However, such outputs are not guaranteed to be valid JSON.
1091
+
> Jais models don't support JSON output formatting (`response_format = { "type": "json_object" }`). You can always prompt the model to generate JSON outputs. However, such outputs are not guaranteed to be valid JSON.
1092
1092
1093
1093
If you want to pass a parameter that isn't in the list of supported parameters, you can pass it to the underlying model using *extra parameters*. See [Pass extra parameters to the model](#pass-extra-parameters-to-the-model).
1094
1094
@@ -1165,14 +1165,14 @@ The following example shows how to handle events when the model detects harmful
1165
1165
1166
1166
## More inference examples
1167
1167
1168
-
For more examples of how to use Jais, see the following examples and tutorials:
1168
+
For more examples of how to use Jais models, see the following examples and tutorials:
## Cost and quota considerations for Jais family ofmodels deployed as serverless API endpoints
1175
+
## Cost and quota considerations for Jais models deployed as serverless API endpoints
1176
1176
1177
1177
Quota is managed per deployment. Each deployment has a rate limit of200,000 tokens per minute and 1,000API requests per minute. However, we currently limit one deployment per model per project. Contact Microsoft Azure Support if the current rate limits aren't sufficient for your scenarios.
1178
1178
@@ -1189,4 +1189,4 @@ For more information on how to track costs, see [Monitor costs for models offere
1189
1189
* [Deploy models as serverless APIs](deploy-models-serverless.md)
1190
1190
* [Consume serverless API endpoints from a different Azure AI Studio project or hub](deploy-models-serverless-connect.md)
1191
1191
* [Region availability for models in serverless API endpoints](deploy-models-serverless-availability.md)
1192
-
* [Plan and manage costs (marketplace)](costs-plan-manage.md#monitor-costs-for-models-offered-through-the-azure-marketplace)
1192
+
* [Plan and manage costs (marketplace)](costs-plan-manage.md#monitor-costs-for-models-offered-through-the-azure-marketplace)
0 commit comments