Skip to content

Commit 33ded2a

Browse files
Merge pull request #7237 from mrbullwinkle/mrb_09_23_2025_PM_requested_updates
[Azure OpenAI] Guidance updates
2 parents 6143690 + a1299f0 commit 33ded2a

File tree

4 files changed

+13
-12
lines changed

4 files changed

+13
-12
lines changed

articles/ai-foundry/openai/concepts/prompt-engineering.md

Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ titleSuffix: Azure OpenAI
44
description: Learn how to use prompt engineering to optimize your work with Azure OpenAI.
55
ms.service: azure-ai-openai
66
ms.topic: conceptual
7-
ms.date: 07/02/2025
7+
ms.date: 09/23/2025
88
ms.custom: references_regions, build-2023, build-2023-dataai
99
manager: nitinme
1010
author: mrbullwinkle
@@ -14,11 +14,9 @@ recommendations: false
1414

1515
# Prompt engineering techniques
1616

17-
GPT-3, GPT-3.5, GPT-4, and GPT-4o models from OpenAI are prompt-based. With prompt-based models, the user interacts with the model by entering a text prompt, to which the model responds with a text completion. This completion is the model’s continuation of the input text. These techniques aren't recommended for o-series models.
17+
These techniques aren't recommended for reasoning models like gpt-5 and o-series models.
1818

19-
While these models are extremely powerful, their behavior is also very sensitive to the prompt. This makes prompt construction an important skill to develop.
20-
21-
Prompt construction can be difficult. In practice, the prompt acts to configure the model weights to complete the desired task, but it's more of an art than a science, often requiring experience and intuition to craft a successful prompt. The goal of this article is to help get you started with this learning process. It attempts to capture general concepts and patterns that apply to all GPT models. However it's important to understand that each model behaves differently, so the learnings might not apply equally to all models.
19+
Prompt construction can be difficult. In practice, the prompt acts assist the model complete the desired task, but it's more of an art than a science, often requiring experience and intuition to craft a successful prompt. The goal of this article is to help get you started with this learning process. It attempts to capture general concepts and patterns that apply to all GPT models. However it's important to understand that each model behaves differently, so the learnings might not apply equally to all models.
2220

2321
## Basics
2422

articles/ai-foundry/openai/concepts/use-your-data.md

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -509,15 +509,10 @@ You can also change the model's output by defining a system message. For example
509509

510510
Azure OpenAI On Your Data works by sending instructions to a large language model in the form of prompts to answer user queries using your data. If there is a certain behavior that is critical to the application, you can repeat the behavior in system message to increase its accuracy. For example, to guide the model to only answer from documents, you can add "*Please answer using retrieved documents only, and without using your knowledge. Please generate citations to retrieved documents for every claim in your answer. If the user question cannot be answered using retrieved documents, please explain the reasoning behind why documents are relevant to user queries. In any case, don't answer using your own knowledge."*
511511

512-
**Prompt Engineering tricks**
513-
514-
There are many tricks in prompt engineering that you can try to improve the output. One example is chain-of-thought prompting where you can add *"Let’s think step by step about information in retrieved documents to answer user queries. Extract relevant knowledge to user queries from documents step by step and form an answer bottom up from the extracted information from relevant documents."*
515-
516512
> [!NOTE]
517513
> The system message is used to modify how GPT assistant responds to a user question based on retrieved documentation. It doesn't affect the retrieval process. If you'd like to provide instructions for the retrieval process, it is better to include them in the questions.
518514
> The system message is only guidance. The model might not adhere to every instruction specified because it has been primed with certain behaviors such as objectivity, and avoiding controversial statements. Unexpected behavior might occur if the system message contradicts with these behaviors.
519515
520-
521516
### Limit responses to your data
522517

523518
This option encourages the model to respond using your data only, and is selected by default. If you unselect this option, the model might more readily apply its internal knowledge to respond. Determine the correct selection based on your use case and scenario.

articles/ai-foundry/openai/how-to/reasoning.md

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -443,8 +443,9 @@ Console.WriteLine($"[ASSISTANT]: {completion.Content[0].Text}");
443443

444444
When using the latest reasoning models with the [Responses API](./responses.md) you can use the reasoning summary parameter to receive summaries of the model's chain of thought reasoning.
445445

446-
> [!NOTE]
447-
> Even when enabled, reasoning summaries are not guaranteed to be generated for every step/request. This is expected behavior.
446+
> [!IMPORTANT]
447+
> Attempting to extract raw reasoning through methods other than the reasoning summary parameter are not supported, may violate the Acceptable Use Policy, and may result in throttling or suspension when detected.
448+
448449

449450
# [Python](#tab/py)
450451

@@ -570,6 +571,9 @@ curl -X POST "https://YOUR-RESOURCE-NAME.openai.azure.com/openai/v1/responses" \
570571
}
571572
```
572573

574+
> [!NOTE]
575+
> Even when enabled, reasoning summaries are not guaranteed to be generated for every step/request. This is expected behavior.
576+
573577
## Python lark
574578

575579
GPT-5 series reasoning models have the ability to call a new `custom_tool` called `lark_tool`. This tool is based on [Python lark](https://github.com/lark-parser/lark) and can be used for more flexible constraining of model output.

articles/ai-foundry/openai/includes/prompt-chat-completion.md

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -98,6 +98,10 @@ One simple way to use an affordance is to stop generation once the affordance ca
9898

9999
## Chain of thought prompting
100100

101+
> [!IMPORTANT]
102+
> This technique is only applicable non-reasoning models. Attempting to extract model reasoning through methods other than the reasoning summary parameter are not supported, may violate the Acceptable Use Policy, and may result in throttling or suspension when detected.
103+
104+
101105
This is a variation on the **break the task down** technique. Instead of splitting a task into smaller steps, in this approach, the model response is instructed to proceed step-by-step and present all the steps involved. Doing so reduces the possibility of inaccuracy of outcomes and makes assessing the model response easier.
102106

103107
| System message |User | Assistant |

0 commit comments

Comments
 (0)