Skip to content

Commit b40325c

Browse files
Merge pull request #6434 from mrbullwinkle/mrb_08_06_2025_vnext_004
[Azure OpenAI] [Release Branch] Update 004
2 parents 18d1672 + 2d5a32b commit b40325c

File tree

1 file changed

+29
-7
lines changed

1 file changed

+29
-7
lines changed

articles/ai-foundry/openai/how-to/reasoning.md

Lines changed: 29 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -37,18 +37,40 @@ Azure OpenAI reasoning models are designed to tackle reasoning and problem-solvi
3737
| `o3` | [Model availability](../concepts/models.md#global-standard-model-availability) | Request access: [o3 limited access model application](https://aka.ms/oai/o3access) |
3838
| `o3-mini` | [Model availability](../concepts/models.md#global-standard-model-availability). | Access is no longer restricted for this model. |
3939
|`o1` | [Model availability](../concepts/models.md#global-standard-model-availability). | Access is no longer restricted for this model. |
40-
| `o1-preview` | [Model availability](../concepts/models.md#global-standard-model-availability). |This model is only available for customers who were granted access as part of the original limited access release. We're currently not expanding access to `o1-preview`. |
4140
| `o1-mini` | [Model availability](../concepts/models.md#global-standard-model-availability). | No access request needed for Global Standard deployments.<br><br>Standard (regional) deployments are currently only available to select customers who were previously granted access as part of the `o1-preview` release.|
4241

4342
## API & feature support
4443

4544
# [GPT-5 Reasoning Models](#tab/gpt-5)
4645

46+
47+
| **Feature** | **gpt-5**, **2025-08-07** | **gpt-5-mini**, **2025-08-07** | **gpt-5-nano**, **2025-08-07** |
48+
|:-------------------|:--------------------------:|:------:|:--------:|
49+
| **API Version** | [v1 preview](../api-version-lifecycle.md#api-evolution) | [v1 preview](../api-version-lifecycle.md#api-evolution) | [v1 preview](../api-version-lifecycle.md#api-evolution) |
50+
| **[Developer Messages](#developer-messages)** ||||
51+
| **[Structured Outputs](./structured-outputs.md)** ||||
52+
| **[Context Window](../concepts/models.md#o-series-models)** | Input: 272,000 <br> Output: 128,000 | Input: 272,000 <br> Output: 128,000 | Input: 272,000 <br> Output: 128,000 |
53+
| **[Reasoning effort](#reasoning-effort)** ||||
54+
| **[Image input](./gpt-with-vision.md)** ||||
55+
| Chat Completions API ||||
56+
| Responses API ||||
57+
| Functions/Tools ||||
58+
| Parallel Tool Calls<sup>1</sup> ||||
59+
| `max_completion_tokens` <sup>2</sup> ||||
60+
| System Messages <sup>3</sup> ||||
61+
| [Reasoning summary](#reasoning-summary) <sup>4</sup> ||||
62+
| Streaming ||||
63+
64+
<sup>1</sup> Parallel tool calls are not supported when `reasoning_effort` is set to `minimal`<br><br>
65+
<sup>2</sup> Reasoning models will only work with the `max_completion_tokens` parameter. <br><br>
66+
<sup>3</sup> The latest reasoning models support system messages to make migration easier. You should not use both a developer message and a system message in the same API request.<br><br>
67+
<sup>4</sup> Access to the chain-of-thought reasoning summary is limited access only for `o3` & `o4-mini`.
68+
4769
# [O-Series Reasoning Models](#tab/o-series)
4870

4971
| **Feature** | **codex-mini**, **2025-05-16** | **o3-pro**, **2025-06-10** | **o4-mini**, **2025-04-16** | **o3**, **2025-04-16** | **o3-mini**, **2025-01-31** |**o1**, **2024-12-17** | **o1-mini**, **2024-09-12** |
5072
|:-------------------|:--------------------------:|:------:|:--------|:-----:|:-------:|:--------------------------:|:---:|
51-
| **API Version** | `2025-04-01-preview` & [v1 preview](../api-version-lifecycle.md#api-evolution) | `2025-04-01-preview` & [v1 preview](../api-version-lifecycle.md#api-evolution) | `2025-04-01-preview` & [v1 preview](../api-version-lifecycle.md#api-evolution) | `2025-04-01-preview` & [v1 preview](../api-version-lifecycle.md#api-evolution) | `2025-04-01-preview` & [v1 preview](../api-version-lifecycle.md#api-evolution) | `2025-04-01-preview` & [v1 preview](../api-version-lifecycle.md#api-evolution) | | `2025-04-01-preview` & [v1 preview](../api-version-lifecycle.md#api-evolution) |
73+
| **API Version** | `2025-04-01-preview` & [v1 preview](../api-version-lifecycle.md#api-evolution) | `2025-04-01-preview` & [v1 preview](../api-version-lifecycle.md#api-evolution) | `2025-04-01-preview` & [v1 preview](../api-version-lifecycle.md#api-evolution) | `2025-04-01-preview` & [v1 preview](../api-version-lifecycle.md#api-evolution) | `2025-04-01-preview` & [v1 preview](../api-version-lifecycle.md#api-evolution) | `2025-04-01-preview` & [v1 preview](../api-version-lifecycle.md#api-evolution) | `2025-04-01-preview` & [v1 preview](../api-version-lifecycle.md#api-evolution) |
5274
| **[Developer Messages](#developer-messages)** ||||||| - |
5375
| **[Structured Outputs](./structured-outputs.md)** ||||||| - |
5476
| **[Context Window](../concepts/models.md#o-series-models)** | Input: 200,000 <br> Output: 100,000 | Input: 200,000 <br> Output: 100,000 | Input: 200,000 <br> Output: 100,000 | Input: 200,000 <br> Output: 100,000 | Input: 200,000 <br> Output: 100,000 | Input: 200,000 <br> Output: 100,000 | Input: 128,000 <br> Output: 65,536 |
@@ -109,7 +131,7 @@ client = AzureOpenAI(
109131
)
110132

111133
response = client.chat.completions.create(
112-
model="o1-new", # replace with the model deployment name of your o1-preview, or o1-mini model
134+
model="o1-new", # replace with your model deployment name
113135
messages=[
114136
{"role": "user", "content": "What steps should I think about when writing my first Python API?"},
115137
],
@@ -312,11 +334,11 @@ token_provider = get_bearer_token_provider(
312334
client = AzureOpenAI(
313335
azure_endpoint = os.getenv("AZURE_OPENAI_ENDPOINT"),
314336
azure_ad_token_provider=token_provider,
315-
api_version="2025-03-01-preview"
337+
api_version="2025-04-01-preview"
316338
)
317339

318340
response = client.chat.completions.create(
319-
model="o1-new", # replace with the model deployment name of your o1-preview, or o1-mini model
341+
model="o1-new", # replace with your model deployment name
320342
messages=[
321343
{"role": "developer","content": "You are a helpful assistant."}, # optional equivalent to a system message for reasoning models
322344
{"role": "user", "content": "What steps should I think about when writing my first Python API?"},
@@ -344,7 +366,7 @@ from openai import AzureOpenAI
344366
client = AzureOpenAI(
345367
azure_endpoint = os.getenv("AZURE_OPENAI_ENDPOINT"),
346368
api_key=os.getenv("AZURE_OPENAI_API_KEY"),
347-
api_version="2025-03-01-preview"
369+
api_version="2025-04-01-preview"
348370
)
349371

350372
response = client.chat.completions.create(
@@ -397,7 +419,7 @@ Console.WriteLine($"{completion.Role}: {completion.Content[0].Text}");
397419

398420
## Reasoning summary
399421

400-
When using the latest `o3` and `o4-mini` models with the [Responses API](./responses.md) you can use the reasoning summary parameter to receive summaries of the model's chain of thought reasoning. This parameter can be set to `auto`, `concise`, or `detailed`. Access to this feature requires you to [Request Access](https://aka.ms/oai/o3access).
422+
When using the latest reasoning models with the [Responses API](./responses.md) you can use the reasoning summary parameter to receive summaries of the model's chain of thought reasoning. This parameter can be set to `auto`, `concise`, or `detailed`. Access to this feature requires you to [Request Access](https://aka.ms/oai/o3access).
401423

402424
> [!NOTE]
403425
> Even when enabled, reasoning summaries are not generated for every step/request. This is expected behavior.

0 commit comments

Comments
 (0)