Skip to content

Commit 3b0d11e

Browse files
authored
Merge pull request #259 from MicrosoftDocs/main
[Publishing] [Scheduled Publish] 9/12/2024 14:10 PM Publish
2 parents 162d4f1 + 4e3e48c commit 3b0d11e

17 files changed

+162
-100
lines changed

articles/ai-services/openai/concepts/models.md

Lines changed: 18 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ titleSuffix: Azure OpenAI
44
description: Learn about the different model capabilities that are available with Azure OpenAI.
55
ms.service: azure-ai-openai
66
ms.topic: conceptual
7-
ms.date: 09/09/2024
7+
ms.date: 09/12/2024
88
ms.custom: references_regions, build-2023, build-2023-dataai, refefences_regions
99
manager: nitinme
1010
author: mrbullwinkle #ChrisHMSFT
@@ -26,6 +26,23 @@ Azure OpenAI Service is powered by a diverse set of models with different capabi
2626
| [Whisper](#whisper-models) | A series of models in preview that can transcribe and translate speech to text. |
2727
| [Text to speech](#text-to-speech-models-preview) (Preview) | A series of models in preview that can synthesize text to speech. |
2828

29+
## o1-preview and o1-mini models limited access
30+
31+
The Azure OpenAI `o1-preview` and `o1-mini` models are specifically designed to tackle reasoning and problem-solving tasks with increased focus and capability. These models spend more time processing and understanding the user's request, making them exceptionally strong in areas like science, coding, and math compared to previous iterations.
32+
33+
### Availability
34+
35+
The `o1-preview` and `o1-mini` models are available in the East US2 region for limited access through the [AI Studio](https://ai.azure.com) early access playground. Data processing for the `o1` models may occur in a different region than where they are available for use.
36+
37+
To try the `o1-preview` and `o1-mini` models in the early access playground, **registration is required, and access will be granted based on Microsoft’s eligibility criteria**.
38+
39+
Request access: [limited access model application](https://aka.ms/oai/modelaccess)
40+
41+
Once access has been granted, you will need to:
42+
43+
1. Navigate to https://ai.azure.com/resources and select a resource in the `eastus2` region. If you do not have an Azure OpenAI resource in this region you will need to [create one](https://portal.azure.com/#create/Microsoft.CognitiveServicesOpenAI).
44+
2. Once the `eastus2` Azure OpenAI resource is selected, in the upper left-hand panel under **Playgrounds** select **Early access playground (preview)**.
45+
2946
## GPT-4o and GPT-4 Turbo
3047

3148
GPT-4o integrates text and images in a single model, enabling it to handle multiple data types simultaneously. This multimodal approach enhances accuracy and responsiveness in human-computer interactions. GPT-4o matches GPT-4 Turbo in English text and coding tasks while offering superior performance in non-English languages and vision tasks, setting new benchmarks for AI capabilities.

articles/ai-services/openai/whats-new.md

Lines changed: 44 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,14 +10,57 @@ ms.custom:
1010
- ignite-2023
1111
- references_regions
1212
ms.topic: whats-new
13-
ms.date: 9/03/2024
13+
ms.date: 9/12/2024
1414
recommendations: false
1515
---
1616

1717
# What's new in Azure OpenAI Service
1818

1919
This article provides a summary of the latest releases and major documentation updates for Azure OpenAI.
2020

21+
## September 2024
22+
23+
### NEW o1-preview and o1-mini models available for limited access
24+
25+
The Azure OpenAI `o1-preview` and `o1-mini` models are specifically designed to tackle reasoning and problem-solving tasks with increased focus and capability. These models spend more time processing and understanding the user's request, making them exceptionally strong in areas like science, coding, and math compared to previous iterations.
26+
27+
### Key capabilities of the o1 series
28+
29+
- Complex Code Generation: Capable of generating algorithms and handling advanced coding tasks to support developers.
30+
- Advanced Problem Solving: Ideal for comprehensive brainstorming sessions and addressing multifaceted challenges.
31+
- Complex Document Comparison: Perfect for analyzing contracts, case files, or legal documents to identify subtle differences.
32+
- Instruction Following and Workflow Management: Particularly effective for managing workflows requiring shorter contexts.
33+
34+
### Model variants
35+
36+
- `o1-preview`: `o1-preview` is the more capable of the `o1` series models.
37+
- `o1-mini`: `o1-mini` is the faster and cheaper of the `o1` series models.
38+
39+
Model version: `2024-09-12`
40+
41+
Request access: [limited access model application](https://aka.ms/oai/modelaccess)
42+
43+
### Limitations
44+
45+
The `o1-preview` and `o1-mini` models are currently in preview and do not include some features available in other models, such as image understanding and structured outputs found in the GPT-4o and GPT-4o-mini models. For many tasks, the generally available GPT-4o models may still be more suitable.
46+
47+
### Safety
48+
49+
OpenAI has incorporated additional safety measures into the `o1` models, including new techniques to help the models refuse unsafe requests. These advancements make the `o1` series some of the most robust models available.
50+
51+
### Availability
52+
53+
The `o1-preview` and `o1-mini` are available in the East US2 region for limited access through the [AI Studio](https://ai.azure.com) early access playground. Data processing for the `o1` models may occur in a different region than where they are available for use.
54+
55+
To try the `o1-preview` and `o1-mini` models in the early access playground **registration is required, and access will be granted based on Microsoft’s eligibility criteria.**
56+
57+
Request access: [limited access model application](https://aka.ms/oai/modelaccess)
58+
59+
Once access has been granted, you will need to:
60+
61+
1. Navigate to https://ai.azure.com/resources and select a resource in the `eastus2` region. If you do not have an Azure OpenAI resource in this region you will need to [create one](https://portal.azure.com/#create/Microsoft.CognitiveServicesOpenAI).
62+
2. Once the `eastus2` Azure OpenAI resource is selected, in the upper left-hand panel under **Playgrounds** select **Early access playground (preview)**.
63+
2164
## August 2024
2265

2366
### GPT-4o 2024-08-06 structured outputs

articles/ai-studio/concepts/rbac-ai-studio.md

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ ms.custom:
88
- ignite-2023
99
- build-2024
1010
ms.topic: conceptual
11-
ms.date: 5/21/2024
11+
ms.date: 9/12/2024
1212
ms.reviewer: deeikele
1313
ms.author: larryfr
1414
author: Blackmist
@@ -220,20 +220,20 @@ When you create a connection that uses Microsoft Entra ID authentication, you mu
220220

221221
| Resource connection | Role | Description |
222222
|----------|------|-------------|
223-
| Azure AI Search | Contributor | List API-Keys to list indexes from Azure OpenAI Studio. |
223+
| Azure AI Search | Contributor | List API-Keys to list indexes from Azure AI Studio. |
224224
| Azure AI Search | Search Index Data Contributor | Required for indexing scenarios |
225-
| Azure AI services/Azure OpenAI | Cognitive Services OpenAI Contributor | Call public ingestion API from Azure OpenAI Studio. |
226-
| Azure AI services/OpenAI | Cognitive Services User | List API-Keys from Azure OpenAI Studio. |
227-
| Azure AI services/OpenAI | Contributor | Allows for calls to the control plane. |
225+
| Azure AI services / Azure OpenAI | Cognitive Services OpenAI Contributor | Call public ingestion API from Azure AI Studio. |
226+
| Azure AI services / Azure OpenAI | Cognitive Services User | List API-Keys from Azure AI Studio. |
227+
| Azure AI services / Azure OpenAI | Contributor | Allows for calls to the control plane. |
228228

229229
When using Microsoft Entra ID authenticated connections in the chat playground, the services need to authorize each other to access the required resources. The admin performing the configuration needs to have the __Owner__ role on these resources to add role assignments. The following table lists the required role assignments for each resource. The __Assignee__ column refers to the system-assigned managed identity of the listed resource. The __Resource__ column refers to the resource that the assignee needs to access. For example, Azure OpenAI has a system-assigned managed identity that needs to be assigned the __Search Index Data Reader__ role for the Azure AI Search resource.
230230

231231
| Role | Assignee | Resource | Description |
232232
|------|----------|----------|-------------|
233-
| Search Index Data Reader | Azure AI services/OpenAI | Azure AI Search | Inference service queries the data from the index. Only used for inference scenarios. |
234-
| Search Index Data Contributor | Azure AI services/OpenAI | Azure AI Search | Read-write access to content in indexes. Import, refresh, or query the documents collection of an index. Only used for ingestion and inference scenarios. |
235-
| Search Service Contributor | Azure AI services/OpenAI | Azure AI Search | Read-write access to object definitions (indexes, aliases, synonym maps, indexers, data sources, and skillsets). Inference service queries the index schema for auto fields mapping. Data ingestion service creates index, data sources, skill set, indexer, and queries the indexer status. |
236-
| Cognitive Services OpenAI Contributor | Azure AI Search | Azure AI services/OpenAI | Custom skill |
233+
| Search Index Data Reader | Azure AI services / Azure OpenAI | Azure AI Search | Inference service queries the data from the index. Only used for inference scenarios. |
234+
| Search Index Data Contributor | Azure AI services / Azure OpenAI | Azure AI Search | Read-write access to content in indexes. Import, refresh, or query the documents collection of an index. Only used for ingestion and inference scenarios. |
235+
| Search Service Contributor | Azure AI services / Azure OpenAI | Azure AI Search | Read-write access to object definitions (indexes, aliases, synonym maps, indexers, data sources, and skillsets). Inference service queries the index schema for auto fields mapping. Data ingestion service creates index, data sources, skill set, indexer, and queries the indexer status. |
236+
| Cognitive Services OpenAI Contributor | Azure AI Search | Azure AI services / Azure OpenAI | Custom skill |
237237
| Cognitive Services OpenAI User | Azure OpenAI Resource for chat model | Azure OpenAI resource for embedding model | Required only if using two Azure OpenAI resources to communicate. |
238238

239239
> [!NOTE]
@@ -316,8 +316,8 @@ The following example defines a role for a developer using [Azure OpenAI Assista
316316
{
317317
"id": "",
318318
"properties": {
319-
"roleName": "CognitiveServices OpenAI Assistants API Developer",
320-
"description": "Custom role to work with AOAI Assistants API",
319+
"roleName": "Azure OpenAI Assistants API Developer",
320+
"description": "Custom role to work with Azure OpenAI Assistants API",
321321
"assignableScopes": [
322322
"<your-scope>"
323323
],

articles/ai-studio/how-to/deploy-models-jais.md

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -201,7 +201,7 @@ print_stream(result)
201201
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](https://aka.ms/azureai/modelinference).
202202

203203
```python
204-
from azure.ai.inference.models import ChatCompletionsResponseFormatText
204+
from azure.ai.inference.models import ChatCompletionsResponseFormat
205205

206206
response = client.complete(
207207
messages=[
@@ -214,12 +214,12 @@ response = client.complete(
214214
stop=["<|endoftext|>"],
215215
temperature=0,
216216
top_p=1,
217-
response_format=ChatCompletionsResponseFormatText(),
217+
response_format={ "type": ChatCompletionsResponseFormat.TEXT },
218218
)
219219
```
220220

221221
> [!WARNING]
222-
> Jais doesn't support JSON output formatting (`response_format = { "type": "json_object" }`). You can always prompt the model to generate JSON outputs. However, such outputs are not guaranteed to be valid JSON.
222+
> Jais models don't support JSON output formatting (`response_format = { "type": "json_object" }`). You can always prompt the model to generate JSON outputs. However, such outputs are not guaranteed to be valid JSON.
223223
224224
If you want to pass a parameter that isn't in the list of supported parameters, you can pass it to the underlying model using *extra parameters*. See [Pass extra parameters to the model](#pass-extra-parameters-to-the-model).
225225

@@ -482,7 +482,7 @@ var response = await client.path("/chat/completions").post({
482482
```
483483
484484
> [!WARNING]
485-
> Jais doesn't support JSON output formatting (`response_format = { "type": "json_object" }`). You can always prompt the model to generate JSON outputs. However, such outputs are not guaranteed to be valid JSON.
485+
> Jais models don't support JSON output formatting (`response_format = { "type": "json_object" }`). You can always prompt the model to generate JSON outputs. However, such outputs are not guaranteed to be valid JSON.
486486
487487
If you want to pass a parameter that isn't in the list of supported parameters, you can pass it to the underlying model using *extra parameters*. See [Pass extra parameters to the model](#pass-extra-parameters-to-the-model).
488488
@@ -580,7 +580,7 @@ Deployment to a serverless API endpoint doesn't require quota from your subscrip
580580
581581
### The inference package installed
582582
583-
You can consume predictions from this model by using the `Azure.AI.Inference` package from [Nuget](https://www.nuget.org/). To install this package, you need the following prerequisites:
583+
You can consume predictions from this model by using the `Azure.AI.Inference` package from [NuGet](https://www.nuget.org/). To install this package, you need the following prerequisites:
584584
585585
* The endpoint URL. To construct the client library, you need to pass in the endpoint URL. The endpoint URL has the form `https://your-host-name.your-azure-region.inference.ai.azure.com`, where `your-host-name` is your unique model deployment host name and `your-azure-region` is the Azure region where the model is deployed (for example, eastus2).
586586
* Depending on your model deployment and authentication preference, you need either a key to authenticate against the service, or Microsoft Entra ID credentials. The key is a 32-character string.
@@ -606,7 +606,7 @@ using Azure.Identity;
606606
using Azure.AI.Inference;
607607
```
608608

609-
This example also use the following namespaces but you may not always need them:
609+
This example also uses the following namespaces but you may not always need them:
610610

611611

612612
```csharp
@@ -775,7 +775,7 @@ Console.WriteLine($"Response: {response.Value.Choices[0].Message.Content}");
775775
```
776776
777777
> [!WARNING]
778-
> Jais doesn't support JSON output formatting (`response_format = { "type": "json_object" }`). You can always prompt the model to generate JSON outputs. However, such outputs are not guaranteed to be valid JSON.
778+
> Jais models don't support JSON output formatting (`response_format = { "type": "json_object" }`). You can always prompt the model to generate JSON outputs. However, such outputs are not guaranteed to be valid JSON.
779779

780780
If you want to pass a parameter that isn't in the list of supported parameters, you can pass it to the underlying model using *extra parameters*. See [Pass extra parameters to the model](#pass-extra-parameters-to-the-model).
781781
@@ -1088,7 +1088,7 @@ Explore other parameters that you can specify in the inference client. For a ful
10881088
```
10891089

10901090
> [!WARNING]
1091-
> Jais doesn't support JSON output formatting (`response_format = { "type": "json_object" }`). You can always prompt the model to generate JSON outputs. However, such outputs are not guaranteed to be valid JSON.
1091+
> Jais models don't support JSON output formatting (`response_format = { "type": "json_object" }`). You can always prompt the model to generate JSON outputs. However, such outputs are not guaranteed to be valid JSON.
10921092
10931093
If you want to pass a parameter that isn't in the list of supported parameters, you can pass it to the underlying model using *extra parameters*. See [Pass extra parameters to the model](#pass-extra-parameters-to-the-model).
10941094

@@ -1165,14 +1165,14 @@ The following example shows how to handle events when the model detects harmful
11651165

11661166
## More inference examples
11671167

1168-
For more examples of how to use Jais, see the following examples and tutorials:
1168+
For more examples of how to use Jais models, see the following examples and tutorials:
11691169

11701170
| Description | Language | Sample |
11711171
|-------------------------------------------|-------------------|-----------------------------------------------------------------|
11721172
| Azure AI Inference package for JavaScript | JavaScript | [Link](https://aka.ms/azsdk/azure-ai-inference/javascript/samples) |
11731173
| Azure AI Inference package for Python | Python | [Link](https://aka.ms/azsdk/azure-ai-inference/python/samples) |
11741174

1175-
## Cost and quota considerations for Jais family of models deployed as serverless API endpoints
1175+
## Cost and quota considerations for Jais models deployed as serverless API endpoints
11761176

11771177
Quota is managed per deployment. Each deployment has a rate limit of 200,000 tokens per minute and 1,000 API requests per minute. However, we currently limit one deployment per model per project. Contact Microsoft Azure Support if the current rate limits aren't sufficient for your scenarios.
11781178
@@ -1189,4 +1189,4 @@ For more information on how to track costs, see [Monitor costs for models offere
11891189
* [Deploy models as serverless APIs](deploy-models-serverless.md)
11901190
* [Consume serverless API endpoints from a different Azure AI Studio project or hub](deploy-models-serverless-connect.md)
11911191
* [Region availability for models in serverless API endpoints](deploy-models-serverless-availability.md)
1192-
* [Plan and manage costs (marketplace)](costs-plan-manage.md#monitor-costs-for-models-offered-through-the-azure-marketplace)
1192+
* [Plan and manage costs (marketplace)](costs-plan-manage.md#monitor-costs-for-models-offered-through-the-azure-marketplace)

0 commit comments

Comments
 (0)