Skip to content

Commit 9b075bf

Browse files
Merge pull request #7383 from mrbullwinkle/mrb_09_30_2025_freshness_003
[Azure OpenAI] [Freshness 003]
2 parents 746b31a + 1b04638 commit 9b075bf

File tree

7 files changed

+16
-25
lines changed

7 files changed

+16
-25
lines changed

articles/ai-foundry/openai/concepts/abuse-monitoring.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ author: mrbullwinkle
66
ms.author: mbullwin
77
ms.service: azure-ai-openai
88
ms.topic: conceptual
9-
ms.date: 07/02/2025
9+
ms.date: 09/30/2025
1010
ms.custom: template-concept, ignite-2024
1111
manager: nitinme
1212
---

articles/ai-foundry/openai/concepts/advanced-prompt-engineering.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ author: mrbullwinkle
66
ms.author: mbullwin
77
ms.service: azure-ai-openai
88
ms.topic: conceptual
9-
ms.date: 07/02/2025
9+
ms.date: 09/30/2025
1010
manager: nitinme
1111
keywords: ChatGPT, GPT-4, meta prompts, chain of thought
1212
---

articles/ai-foundry/openai/concepts/red-teaming.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ description: Learn about how red teaming and adversarial testing are an essentia
55
author: mrbullwinkle
66
ms.author: mbullwin
77
manager: nitinme
8-
ms.date: 07/02/2025
8+
ms.date: 09/30/2025
99
ms.service: azure-ai-openai
1010
ms.topic: conceptual
1111
ms.custom:

articles/ai-foundry/openai/concepts/system-message.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ titleSuffix: Azure OpenAI in Azure AI Foundry Models
44
description: Learn about how to construct system messages also know as metaprompts to guide an AI system's behavior.
55
ms.service: azure-ai-openai
66
ms.topic: conceptual
7-
ms.date: 07/02/2025
7+
ms.date: 09/30/2025
88
ms.custom:
99
- ignite-2023
1010
manager: nitinme

articles/ai-foundry/openai/concepts/understand-embeddings.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ description: Learn more about how the Azure OpenAI embeddings API uses cosine si
55
manager: nitinme
66
ms.service: azure-ai-openai
77
ms.topic: tutorial
8-
ms.date: 07/02/2025
8+
ms.date: 09/30/2025
99
author: mrbullwinkle
1010
ms.author: mbullwin
1111
recommendations: false

articles/ai-foundry/openai/includes/api-versions/latest-inference-preview.md

Lines changed: 9 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -4647,7 +4647,7 @@ Creates a model response.
46474647
| max_output_tokens | integer | An upper bound for the number of tokens that can be generated for a response, including visible output tokens and conversation state.<br> | No | |
46484648
| parallel_tool_calls | boolean | Whether to allow the model to run tool calls in parallel.<br> | No | True |
46494649
| previous_response_id | string | The unique ID of the previous response to the model. Use this to create multi-turn conversations. Learn more about conversation state.<br> | No | |
4650-
| reasoning | [Reasoning](#reasoning) | **o-series models only**<br><br>Configuration options for reasoning models.<br> | No | |
4650+
| reasoning | [Reasoning](#reasoning) |Configuration options for reasoning models. | No | |
46514651
| store | boolean | Whether to store the generated model response for later retrieval via API.<br> | No | True |
46524652
| stream | boolean | If set to true, the model response data will be streamed to the client as it is generated using [server-sent events](https://developer.mozilla.org/docs/Web/API/Server-sent_events/Using_server-sent_events#Event_stream_format).<br>See the Streaming section below for more information.<br> | No | False |
46534653
| text | object | Configuration options for a text response from the model. Can be plain text or structured JSON data. Learn more:<br>- Text inputs and outputs<br>- Structured Outputs | No | |
@@ -8408,7 +8408,7 @@ An x/y coordinate pair, e.g. `{ x: 100, y: 200 }`.
84088408
| max_output_tokens | integer | An upper bound for the number of tokens that can be generated for a response, including visible output tokens and reasoning tokens.<br> | No | |
84098409
| parallel_tool_calls | boolean | Whether to allow the model to run tool calls in parallel.<br> | No | True |
84108410
| previous_response_id | string | The unique ID of the previous response to the model. Use this to create multi-turn conversations. | No | |
8411-
| reasoning | [Reasoning](#reasoning) | **o-series models only**<br><br>Configuration options for reasoning models.<br> | No | |
8411+
| reasoning | [Reasoning](#reasoning) | Configuration options for reasoning models. | No | |
84128412
| store | boolean | Whether to store the generated model response for later retrieval via API.<br> | No | True |
84138413
| stream | boolean | If set to true, the model response data will be streamed to the client as it is generated using [server-sent events](https://developer.mozilla.org/docs/Web/API/Server-sent_events/Using_server-sent_events#Event_stream_format). | No | False |
84148414
| text | object | Configuration options for a text response from the model. Can be plain text or structured JSON data. Learn more:<br>- text inputs and outputs<br>- Structured Outputs<br> | No | |
@@ -8922,16 +8922,13 @@ When a session is created on the server via REST API, the session object also co
89228922

89238923
### Reasoning
89248924

8925-
**o-series models only**
8926-
8927-
Configuration options for
8928-
reasoning models.
8925+
Configuration options for reasoning models.
89298926

89308927

89318928
| Name | Type | Description | Required | Default |
89328929
|------|------|-------------|----------|---------|
8933-
| effort | [ReasoningEffort](#reasoningeffort) | **o-series models only** <br><br>Constrains effort on reasoning for reasoning models.<br>Currently supported values are `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.<br> | Yes | medium |
8934-
| summary | enum | **o-series models only** <br><br>A summary of the reasoning performed by the model. This can be useful for debugging and understanding the model's reasoning process.<br>One of `concise` or `detailed`.<br><br>Possible values: `concise`, `detailed` | No | |
8930+
| effort | [ReasoningEffort](#reasoningeffort) | Constrains effort on reasoning for reasoning models.<br>Currently supported values are `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.<br> | Yes | medium |
8931+
| summary | enum | A summary of the reasoning performed by the model. This can be useful for debugging and understanding the model's reasoning process.<br>One of `concise` or `detailed`.<br><br>Possible values: `concise`, `detailed` | No | |
89358932

89368933
### ReasoningItem
89378934

@@ -8974,7 +8971,7 @@ A refusal from the model.
89748971
| output_text | string | SDK-only convenience property that contains the aggregated text output from all `output_text` items in the `output` array, if any are present. <br>Supported in the Python and JavaScript SDKs.<br> | No | |
89758972
| parallel_tool_calls | boolean | Whether to allow the model to run tool calls in parallel.<br> | Yes | True |
89768973
| previous_response_id | string | The unique ID of the previous response to the model. Use this to create multi-turn conversations. | No | |
8977-
| reasoning | [Reasoning](#reasoning) | **o-series models only**<br><br>Configuration options for reasoning models.<br> | No | |
8974+
| reasoning | [Reasoning](#reasoning) | Configuration options for reasoning models.<br> | No | |
89788975
| status | enum | The status of the response generation. One of `completed`, `failed`, `in_progress`, or `incomplete`.<br><br>Possible values: `completed`, `failed`, `in_progress`, `incomplete` | No | |
89798976
| temperature | number | What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic.<br>We generally recommend altering this or `top_p` but not both.<br> | Yes | 1 |
89808977
| text | object | Configuration options for a text response from the model. Can be plain text or structured JSON data. Learn more:<br>- text inputs and outputs<br>- Structured Outputs<br> | No | |
@@ -9295,7 +9292,7 @@ Emitted when an output item is marked done.
92959292
| instructions | string | Inserts a system (or developer) message as the first item in the model's context.<br><br>When using along with `previous_response_id`, the instructions from a previous response will be not be carried over to the next response. This makes it simple to swap out system (or developer) messages in new responses.<br> | No | |
92969293
| max_output_tokens | integer | An upper bound for the number of tokens that can be generated for a response, including visible output tokens and conversation state.<br> | No | |
92979294
| previous_response_id | string | The unique ID of the previous response to the model. Use this to create multi-turn conversations. | No | |
9298-
| reasoning | [Reasoning](#reasoning) | **o-series models only**<br><br>Configuration options for reasoning models.<br> | No | |
9295+
| reasoning | [Reasoning](#reasoning) | Configuration options for reasoning models.<br> | No | |
92999296
| text | object | Configuration options for a text response from the model. Can be plain text or structured JSON data. Learn more:<br>- text inputs and outputs<br>- Structured Outputs<br> | No | |
93009297
| └─ format | [TextResponseFormatConfiguration](#textresponseformatconfiguration) | An object specifying the format that the model must output.<br><br>Configuring `{ "type": "json_schema" }` enables Structured Outputs, which ensures the model matches your supplied JSON schema. The default format is `{ "type": "text" }` with no additional options.<br><br>**Not recommended for gpt-4o and newer models:**<br><br>Setting to `{ "type": "json_object" }` enables the older JSON mode, which ensures the message the model generates is valid JSON. Using `json_schema` is preferred for models that support it.<br> | No | |
93019298
| tool_choice | [ToolChoiceOptions](#toolchoiceoptions) or [ToolChoiceTypes](#toolchoicetypes) or [ToolChoiceFunction](#toolchoicefunction) | How the model should select which tool (or tools) to use when generating a response. See the `tools` parameter to see how to specify which tools the model can call.<br> | No | |
@@ -9618,18 +9615,12 @@ A wait action.
96189615

96199616
### ReasoningEffort
96209617

9621-
**o-series models only**
9622-
9623-
Constrains effort on reasoning for
9624-
reasoning models.
9625-
Currently supported values are `low`, `medium`, and `high`. Reducing
9626-
reasoning effort can result in faster responses and fewer tokens used
9627-
on reasoning in a response.
9618+
Constrains effort on reasoning for reasoning models. Currently supported values are `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.
96289619

96299620

96309621
| Property | Value |
96319622
|----------|-------|
9632-
| **Description** | **o-series models only** <br><br>Constrains effort on reasoning for reasoning models.<br>Currently supported values are `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.<br> |
9623+
| **Description** | Constrains effort on reasoning for reasoning models.<br>Currently supported values are `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.<br> |
96339624
| **Type** | string |
96349625
| **Default** | medium |
96359626
| **Values** | `low`<br>`medium`<br>`high` |

articles/ai-foundry/openai/reference-preview.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ description: Learn how to use Azure OpenAI's latest preview REST API. In this ar
55
manager: nitinme
66
ms.service: azure-ai-openai
77
ms.topic: conceptual
8-
ms.date: 07/02/2025
8+
ms.date: 09/30/2025
99
author: mrbullwinkle
1010
ms.author: mbullwin
1111
recommendations: false
@@ -22,7 +22,7 @@ This article provides details on the inference REST API endpoints for Azure Open
2222

2323
The rest of the article covers the `2025-04-01-preview` preview release of the Azure OpenAI data plane inference specification.
2424

25-
If you're looking for documentation on the latest GA API release, refer to the [latest GA data plane inference API](./reference.md)
25+
If you're looking for documentation on the latest GA API release, refer to the [latest GA data plane inference API](./latest.md)
2626

2727
[!INCLUDE [API surfaces](./includes/api-versions/latest-inference-preview.md)]
2828

0 commit comments

Comments
 (0)