Skip to content

Commit edea101

Browse files
committed
update
1 parent 7509cb2 commit edea101

File tree

1 file changed

+19
-9
lines changed

1 file changed

+19
-9
lines changed

articles/ai-services/openai/how-to/reasoning.md

Lines changed: 19 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ description: Learn how to use Azure OpenAI's advanced o1 series reasoning models
55
manager: nitinme
66
ms.service: azure-ai-openai
77
ms.topic: include
8-
ms.date: 12/17/2024
8+
ms.date: 01/16/2025
99
author: mrbullwinkle
1010
ms.author: mbullwin
1111
---
@@ -38,21 +38,31 @@ Once access has been granted, you'll need to create a deployment for each model.
3838
| `o1-preview` | See [models page](../concepts/models.md#global-standard-model-availability). |
3939
| `o1-mini` | See [models page](../concepts/models.md#global-standard-model-availability). |
4040

41-
## API support
41+
## API & feature support
4242

43-
Initial support for the **o1-preview** and **o1-mini** preview models was added in API version `2024-09-01-preview`.
43+
| **Feature** | **o1**, **2024-12-17** | **o1-preview**, **2024-09-12** | **o1-mini**, **2024-09-12** |
44+
|:-------------------|:--------------------------:|:--------------------------:|:-------------------------------:|
45+
| **API Version** | `2024-12-01-preview` | `2024-09-01-preview` <br> `2024-10-01-preview` <br> `2024-12-01-preview` | `2024-09-01-preview` <br> `2024-10-01-preview` <br> `2024-12-01-preview` |
46+
| **[Developer Messages](#developer-messages)** || - | - |
47+
| **[Structured Outputs](./structured-outputs.md)** || - | - |
48+
| **[Context Window](../concepts/models.md#o1-and-o1-mini-models-limited-access)** | Input: 200,000 <br> Output: 100,000 | Input: 128,000 <br> Output: 32,768 | Input: 128,000 <br> Output: 65,536 |
49+
| **[Reasoning effort](#reasoning-effort)** || - | - |
50+
| System Messages | - | - | - |
4451

45-
As part of this release, the `max_tokens` parameter was deprecated and replaced with the new `max_completion_tokens` parameter. **o1 series** models will only work with the `max_completion_tokens` parameter.
52+
`max_tokens` parameter has been replaced with the new `max_completion_tokens` parameter. **o1 series** models will only work with the `max_completion_tokens` parameter.
4653

47-
The latest most capable **o1 series** model is `o1` **Version: 2024-12-17**. This general availability (GA) model should be used with API version `2024-12-01-preview`.
54+
### Not Supported
4855

49-
### 2024-12-01-preview
56+
The following are currently unsupported with o1-series models:
5057

51-
`2024-12-01-preview` adds support for the new `reasoning_effort` parameter, [structured outputs](./structured-outputs.md), and developer messages. The older preview reasoning models do not currently support these features. For reasoning models, these features are currently only available with `o1` **Version: 2024-12-17**.
58+
- System Messages
59+
- Streaming
60+
- Parallel tool calling
61+
- `temperature`, `top_p`, `presence_penalty`, `frequency_penalty`, `logprobs`, `top_logprobs`, `logit_bias`
5262

5363
## Usage
5464

55-
These models do not currently support the same set of parameters as other models that use the chat completions API. Only a limited subset is currently supported. Using standard parameters like `temperature` and `top_p` will result in errors.
65+
These models do [not currently support](#api--feature-support) the same set of parameters as other models that use the chat completions API.
5666

5767
# [Python (Microsoft Entra ID)](#tab/python-secure)
5868

@@ -218,7 +228,7 @@ print(response.model_dump_json(indent=2))
218228
}
219229
```
220230

221-
231+
## Reasoning effort
222232

223233
> [!NOTE]
224234
> Reasoning models have `reasoning_tokens` as part of `completion_tokens_details` in the model response. These are hidden tokens that are not returned as part of the message response content but are used by the model to help generate a final answer to your request. `2024-12-01-preview` adds an additional new parameter `reasoning_effort` which can be set to `low`, `medium`, or `high` with the latest `o1` model. The higher the effort setting, the longer the model will spend processing the request, which will generally result in a larger number of `reasoning_tokens`.

0 commit comments

Comments
 (0)