You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-foundry/agents/how-to/tools/model-context-protocol.md
+1-5Lines changed: 1 addition & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -7,17 +7,13 @@ manager: nitinme
7
7
ms.service: azure-ai-foundry
8
8
ms.subservice: azure-ai-foundry-agent-service
9
9
ms.topic: how-to
10
-
ms.date: 09/04/2025
10
+
ms.date: 09/30/2025
11
11
author: aahill
12
12
ms.author: aahi
13
-
ms.custom: references_regions
14
13
---
15
14
16
15
# Connect to Model Context Protocol servers (preview)
17
16
18
-
> [!NOTE]
19
-
> Supported regions are `westus`, `westus2`, `uaenorth`, `southindia`, and `switzerlandnorth`.
20
-
21
17
> [!NOTE]
22
18
> When using a [Network Secured Azure AI Foundry](../../how-to/virtual-networks.md), private MCP servers deployed in the same virtual network is not supported, only publicly accessible MCP servers are supported.
@@ -102,27 +102,10 @@ Supporting content is information that the model can utilize to influence the ou
102
102
103
103
## Scenario-specific guidance
104
104
105
-
While the principles of prompt engineering can be generalized across many different model types, certain models expect a specialized prompt structure. For Azure OpenAI GPT models, there are currently two distinct APIs where prompt engineering comes into play:
106
-
107
-
- Chat Completion API.
108
-
- Completion API.
109
-
110
-
Each API requires input data to be formatted differently, which in turn impacts overall prompt design. The **Chat Completion API** supports the GPT-35-Turbo and GPT-4 models. These models are designed to take input formatted in a [specific chat-like transcript](../how-to/chatgpt.md) stored inside an array of dictionaries.
111
-
112
-
The **Completion API** supports the older GPT-3 models and has much more flexible input requirements in that it takes a string of text with no specific format rules.
113
-
114
105
The techniques in this section will teach you strategies for increasing the accuracy and grounding of responses you generate with a Large Language Model (LLM). It is, however, important to remember that even when using prompt engineering effectively you still need to validate the responses the models generate. Just because a carefully crafted prompt worked well for a particular scenario doesn't necessarily mean it will generalize more broadly to certain use cases. Understanding the [limitations of LLMs](/azure/ai-foundry/responsible-ai/openai/transparency-note#limitations), is just as important as understanding how to leverage their strengths.
-**Be Specific**. Leave as little to interpretation as possible. Restrict the operational space.
@@ -133,7 +116,7 @@ The techniques in this section will teach you strategies for increasing the accu
133
116
134
117
## Space efficiency
135
118
136
-
While the input size increases with each new generation of GPT models, there will continue to be scenarios that provide more data than the model can handle. GPT models break words into "tokens." While common multi-syllable words are often a single token, less common words are broken in syllables. Tokens can sometimes be counter-intuitive, as shown by the example below which demonstrates token boundaries for different date formats. In this case, spelling out the entire month is more space efficient than a fully numeric date. The current range of token support goes from 2,000 tokens with earlier GPT-3 models to up to 32,768 tokens with the 32k version of the latest GPT-4 model.
119
+
While the input size increases with each new generation of GPT models, there will continue to be scenarios that provide more data than the model can handle. GPT models break words into "tokens." While common multi-syllable words are often a single token, less common words are broken in syllables. Tokens can sometimes be counter-intuitive, as shown by the example below which demonstrates token boundaries for different date formats. In this case, spelling out the entire month is more space efficient than a fully numeric date.
137
120
138
121
:::image type="content" source="../media/prompt-engineering/space-efficiency.png" alt-text="Screenshot of a string of text with highlighted colors delineating token boundaries." lightbox="../media/prompt-engineering/space-efficiency.png":::
Copy file name to clipboardExpand all lines: articles/ai-foundry/openai/how-to/function-calling.md
+2-4Lines changed: 2 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -7,7 +7,7 @@ ms.author: mbullwin #delegenz
7
7
ms.service: azure-ai-openai
8
8
ms.custom: devx-track-python
9
9
ms.topic: how-to
10
-
ms.date: 09/15/2025
10
+
ms.date: 09/30/2025
11
11
manager: nitinme
12
12
---
13
13
@@ -31,9 +31,6 @@ At a high level you can break down working with functions into three steps:
31
31
32
32
*`gpt-35-turbo` (`1106`)
33
33
*`gpt-35-turbo` (`0125`)
34
-
*`gpt-4` (`1106-Preview`)
35
-
*`gpt-4` (`0125-Preview`)
36
-
*`gpt-4` (`vision-preview`)
37
34
*`gpt-4` (`2024-04-09`)
38
35
*`gpt-4o` (`2024-05-13`)
39
36
*`gpt-4o` (`2024-08-06`)
@@ -44,6 +41,7 @@ At a high level you can break down working with functions into three steps:
44
41
*`gpt-5` (`2025-08-07`)
45
42
*`gpt-5-mini` (`2025-08-07`)
46
43
*`gpt-5-nano` (`2025-08-07`)
44
+
*`gpt-5-codex` (`2025-09-11`)
47
45
48
46
Support for parallel function was first added in API version [`2023-12-01-preview`](https://github.com/Azure/azure-rest-api-specs/blob/main/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2023-12-01-preview/inference.json)
0 commit comments