Skip to content

Commit 60e9bf7

Browse files
committed
update
2 parents 78977a2 + fff34e3 commit 60e9bf7

File tree

9 files changed

+51
-44
lines changed

9 files changed

+51
-44
lines changed

.openpublishing.publish.config.json

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -117,7 +117,7 @@
117117
"branch_mapping": {}
118118
},
119119
{
120-
"path_to_root": "foundry-samples",
120+
"path_to_root": "foundry-samples-main",
121121
"url": "https://github.com/azure-ai-foundry/foundry-samples",
122122
"branch": "main",
123123
"branch_mapping": {}

articles/ai-foundry/concepts/observability.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -86,7 +86,7 @@ Evaluators are specialized tools that measure the quality, safety, and reliabili
8686
| METEOR | Metric for Evaluation of Translation with Explicit Ordering measures overlaps in n-grams between response and ground truth. |
8787

8888

89-
[**Azure OpenAI Graders (preivew):**](./evaluation-evaluators/azure-openai-graders.md)
89+
[**Azure OpenAI Graders (preview):**](./evaluation-evaluators/azure-openai-graders.md)
9090

9191
| Evaluator | Purpose |
9292
|--|--|

articles/ai-foundry/includes/get-started-fdp.md

Lines changed: 18 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ No installation is necessary to use the Azure AI Foundry portal.
7474
1. [Install Java and Azure CLI](../how-to/develop/install-cli-sdk.md?pivots=programming-language-java).
7575
1. Set these environment variables to use in your scripts:
7676
77-
:::code language="plaintext" source="~/foundry-samples/samples/microsoft/java/mslearn-resources/quickstart/.env.template":::
77+
:::code language="plaintext" source="~/foundry-samples-main/samples/microsoft/java/mslearn-resources/quickstart/.env.template":::
7878
7979
1. Make sure to sign in using the CLI `az login` (or `az login --use-device-code`) command to authenticate before running your Java scripts.
8080
1. Download [POM.XML](https://github.com/azure-ai-foundry/foundry-samples/blob/main/samples/microsoft/java/mslearn-resources/quickstart/pom.xml) to your Java IDE.
@@ -90,7 +90,7 @@ No installation is necessary to use the Azure AI Foundry portal.
9090
1. Install packages with `npm install`
9191
1. Set these environment variables to use in your scripts:
9292
93-
:::code language="plaintext" source="~/foundry-samples/samples/microsoft/javascript/mslearn-resources/quickstart/.env.template":::
93+
:::code language="plaintext" source="~/foundry-samples-main/samples/microsoft/javascript/mslearn-resources/quickstart/.env.template":::
9494
9595
9696
> [!TIP]
@@ -105,7 +105,7 @@ No installation is necessary to use the Azure AI Foundry portal.
105105
106106
1. Set these environment variables to use in your scripts:
107107
108-
:::code language="plaintext" source="~/foundry-samples/samples/microsoft/csharp/mslearn-resources/quickstart/Samples/.env.example":::
108+
:::code language="plaintext" source="~/foundry-samples-main/samples/microsoft/csharp/mslearn-resources/quickstart/Samples/.env.example":::
109109
110110
1. Make sure to sign in using the CLI `az login` (or `az login --use-device-code`) command to authenticate before running your C# scripts.
111111
@@ -146,27 +146,27 @@ Chat completions are the basic building block of AI applications. Using chat com
146146
147147
Substitute your value for the endpoint in this code:
148148
149-
:::code language="python" source="~/foundry-samples/samples/microsoft/python/mslearn-resources/quickstart/quickstart.py" id="chat_completion":::
149+
:::code language="python" source="~/foundry-samples-main/samples/microsoft/python/mslearn-resources/quickstart/quickstart.py" id="chat_completion":::
150150
151151
152152
# [Java (preview)](#tab/java)
153153
154-
:::code language="java" source="~/foundry-samples/samples/microsoft/java/mslearn-resources/quickstart/src/main/java/com/azure/ai/foundry/samples/ChatCompletionSample.java" :::
154+
:::code language="java" source="~/foundry-samples-main/samples/microsoft/java/mslearn-resources/quickstart/src/main/java/com/azure/ai/foundry/samples/ChatCompletionSample.java" :::
155155
156156
# [JavaScript (preview)](#tab/javascript)
157157
158-
:::code language="javascript" source="~/foundry-samples/samples/microsoft/javascript/mslearn-resources/quickstart/src/quickstart.js" id="chat_completion":::
158+
:::code language="javascript" source="~/foundry-samples-main/samples/microsoft/javascript/mslearn-resources/quickstart/src/quickstart.js" id="chat_completion":::
159159
160160
161161
# [C#](#tab/csharp)
162162
163-
:::code language="csharp" source="~/foundry-samples/samples/microsoft/csharp/mslearn-resources/quickstart/Samples/SimpleInference.cs" id="chat_completion":::
163+
:::code language="csharp" source="~/foundry-samples-main/samples/microsoft/csharp/mslearn-resources/quickstart/Samples/SimpleInference.cs" id="chat_completion":::
164164
165165
# [REST API](#tab/rest)
166166
167167
Replace `YOUR-FOUNDRY-RESOURCE-NAME` with your values:
168168
169-
:::code language="console" source="~/foundry-samples/samples/microsoft/REST/mslearn-resources/quickstart/quickstart.sh" id="chat_completion":::
169+
:::code language="console" source="~/foundry-samples-main/samples/microsoft/REST/mslearn-resources/quickstart/quickstart.sh" id="chat_completion":::
170170
171171
---
172172
@@ -186,25 +186,25 @@ Agents allow more powerful capabilities through the use of tools. First, let's w
186186
187187
# [Python](#tab/python)
188188
189-
:::code language="python" source="~/foundry-samples/samples/microsoft/python/mslearn-resources/quickstart/quickstart.py" id="create_and_run_agent":::
189+
:::code language="python" source="~/foundry-samples-main/samples/microsoft/python/mslearn-resources/quickstart/quickstart.py" id="create_and_run_agent":::
190190
191191
# [Java (preview)](#tab/java)
192192
193-
:::code language="java" source="~/foundry-samples/samples/microsoft/java/mslearn-resources/quickstart/src/main/java/com/azure/ai/foundry/samples/AgentSample.java" :::
193+
:::code language="java" source="~/foundry-samples-main/samples/microsoft/java/mslearn-resources/quickstart/src/main/java/com/azure/ai/foundry/samples/AgentSample.java" :::
194194
195195
# [JavaScript (preview)](#tab/javascript)
196196
197-
:::code language="javascript" source="~/foundry-samples/samples/microsoft/javascript/mslearn-resources/quickstart/src/quickstart.js" id="create_and_run_agent" :::
197+
:::code language="javascript" source="~/foundry-samples-main/samples/microsoft/javascript/mslearn-resources/quickstart/src/quickstart.js" id="create_and_run_agent" :::
198198
199199
# [C#](#tab/csharp)
200200
201-
:::code language="csharp" source="~/foundry-samples/samples/microsoft/csharp/mslearn-resources/quickstart/Samples/AgentService.cs" id="create_and_run_agent" :::
201+
:::code language="csharp" source="~/foundry-samples-main/samples/microsoft/csharp/mslearn-resources/quickstart/Samples/AgentService.cs" id="create_and_run_agent" :::
202202
203203
# [REST API](#tab/rest)
204204
205205
Replace `YOUR-FOUNDRY-RESOURCE-NAME` and `YOUR-PROJECT-NAME` with your values:
206206
207-
:::code language="console" source="~/foundry-samples/samples/microsoft/REST/mslearn-resources/quickstart/quickstart.sh" id="create_and_run_agent":::
207+
:::code language="console" source="~/foundry-samples-main/samples/microsoft/REST/mslearn-resources/quickstart/quickstart.sh" id="create_and_run_agent":::
208208
209209
210210
---
@@ -229,26 +229,26 @@ Now let's add a file search tool that enables us to do knowledge retrieval.
229229
230230
# [Python](#tab/python)
231231
232-
:::code language="python" source="~/foundry-samples/samples/microsoft/python/mslearn-resources/quickstart/quickstart.py" id="create_filesearch_agent":::
232+
:::code language="python" source="~/foundry-samples-main/samples/microsoft/python/mslearn-resources/quickstart/quickstart.py" id="create_filesearch_agent":::
233233
234234
# [Java (preview)](#tab/java)
235235
236-
:::code language="java" source="~/foundry-samples/samples/microsoft/java/mslearn-resources/quickstart/src/main/java/com/azure/ai/foundry/samples/FileSearchAgentSample.java" :::
236+
:::code language="java" source="~/foundry-samples-main/samples/microsoft/java/mslearn-resources/quickstart/src/main/java/com/azure/ai/foundry/samples/FileSearchAgentSample.java" :::
237237
238238
239239
# [JavaScript (preview)](#tab/javascript)
240240
241-
:::code language="javascript" source="~/foundry-samples/samples/microsoft/javascript/mslearn-resources/quickstart/src/quickstart.js" id="create_filesearch_agent":::
241+
:::code language="javascript" source="~/foundry-samples-main/samples/microsoft/javascript/mslearn-resources/quickstart/src/quickstart.js" id="create_filesearch_agent":::
242242
243243
# [C#](#tab/csharp)
244244
245-
:::code language="csharp" source="~/foundry-samples/samples/microsoft/csharp/mslearn-resources/quickstart/Samples/AgentFileSearch.cs" id="create_filesearch_agent":::
245+
:::code language="csharp" source="~/foundry-samples-main/samples/microsoft/csharp/mslearn-resources/quickstart/Samples/AgentFileSearch.cs" id="create_filesearch_agent":::
246246
247247
# [REST API](#tab/rest)
248248
249249
Replace `YOUR-FOUNDRY-RESOURCE-NAME` and `YOUR-PROJECT-NAME` with your values:
250250
251-
:::code language="console" source="~/foundry-samples/samples/microsoft/REST/mslearn-resources/quickstart/quickstart.sh" id="create_filesearch_agent":::
251+
:::code language="console" source="~/foundry-samples-main/samples/microsoft/REST/mslearn-resources/quickstart/quickstart.sh" id="create_filesearch_agent":::
252252
253253
---
254254

articles/ai-services/agents/quickstart.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ zone_pivot_groups: agents-quickstart
1212
ms.custom: azure-ai-agents
1313
---
1414

15-
# Quickstart: Create a new agent (Preview)
15+
# Quickstart: Create a new agent
1616

1717
Azure AI Foundry Agent Service allows you to create AI agents tailored to your needs through custom instructions and augmented by advanced tools like code interpreter, and custom functions.
1818

articles/ai-services/openai/concepts/model-router.md

Lines changed: 18 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
---
2-
title: Azure OpenAI model router (preview) concepts
2+
title: Model router for Azure AI Foundry (preview) concepts
33
titleSuffix: Azure OpenAI
44
description: Learn about the model router feature in Azure OpenAI in Azure AI Foundry Models.
55
author: PatrickFarley
@@ -11,13 +11,13 @@ ms.custom:
1111
manager: nitinme
1212
---
1313

14-
# Azure OpenAI model router (preview)
14+
# Model router for Azure AI Foundry (preview)
1515

16-
Azure OpenAI model router is a deployable AI chat model that is trained to select the best large language model (LLM) to respond to a given prompt in real time. By evaluating factors like query complexity, cost, and performance, it intelligently routes requests to the most suitable model.
16+
Model router for Azure AI Foundry is a deployable AI chat model that is trained to select the best large language model (LLM) to respond to a given prompt in real time. By evaluating factors like query complexity, cost, and performance, it intelligently routes requests to the most suitable model. Thus, it delivers high performance while saving on compute costs where possible, all packaged as a single model deployment.
1717

1818
## Why use model router?
1919

20-
Model router intelligently selects the best underlying model for a given prompt to optimize costs while maintaining quality. Smaller and cheaper models are used when they're sufficient for the task, but larger and more expensive models are available for more complex tasks. Also, reasoning models are available for tasks that require complex reasoning, and non-reasoning models are used otherwise. Model router provides a single chat experience that combines the best features from all of the underlying chat models.
20+
Model router intelligently selects the best underlying model for a given prompt to optimize costs while maintaining quality. Smaller and cheaper models are used when they're sufficient for the task, but larger and more expensive models are available for more complex tasks. Also, reasoning models are available for tasks that require complex reasoning, and non-reasoning models are used otherwise. Model router provides a single deployment and chat experience that combines the best features from all of the underlying chat models.
2121

2222
## Versioning
2323

@@ -29,24 +29,32 @@ If you select **Auto-update** at the deployment step (see [Manage models](/azure
2929

3030
|Model router version|Underlying models (version)|
3131
|---|---|
32-
|`2025-04-15`|GPT-4.1 (`2025-04-14`)</br>GPT-4.1-mini (`2025-04-14`)</br>GPT-4.1-nano (`2025-04-14`) </br>o4-mini (`2025-04-16`) |
32+
|`2025-05-19`|GPT-4.1 (`2025-04-14`)</br>GPT-4.1-mini (`2025-04-14`)</br>GPT-4.1-nano (`2025-04-14`) </br>o4-mini (`2025-04-16`) |
3333

3434

3535
## Limitations
3636

37+
### Resource limitations
38+
39+
See the [Models](../concepts/models.md#model-router) page for the region availability and deployment types for model router.
40+
41+
### Technical limitations
42+
3743
See [Quotas and limits](/azure/ai-services/openai/quotas-limits) for rate limit information.
3844

39-
The context window limit listed on the [Models](../concepts/models.md) page is the limit of the smallest underlying model. Other underlying models are compatible with larger context windows, which means an API call with a larger context will succeed only if the prompt happens to be routed to the right model, otherwise the call will fail. To shorten the context window, you can do one of the following:
40-
- Summarize the prompt before passing it to the model
41-
- Truncate the prompt into more relevant parts
42-
- Use document embeddings and have the chat model retrieve relevant sections: see [Azure AI Search](/azure/search/search-what-is-azure-search)
45+
> [!NOTE]
46+
> The context window limit listed on the [Models](../concepts/models.md#model-router) page is the limit of the smallest underlying model. Other underlying models are compatible with larger context windows, which means an API call with a larger context will succeed only if the prompt happens to be routed to the right model, otherwise the call will fail. To shorten the context window, you can do one of the following:
47+
> - Summarize the prompt before passing it to the model
48+
> - Truncate the prompt into more relevant parts
49+
> - Use document embeddings and have the chat model retrieve relevant sections: see [Azure AI Search](/azure/search/search-what-is-azure-search)
4350
51+
Model router accepts image inputs for [Vision enabled chats](/azure/ai-services/openai/how-to/gpt-with-vision) (all of the underlying models can accept image input), but the routing decision is based on the text input only.
4452

4553
Model router doesn't process audio input.
4654

4755
## Billing information
4856

49-
When you use Azure OpenAI model router, you're only billed for the use of the underlying models as they're recruited to respond to prompts. The model router itself doesn't incur any extra charges.
57+
When you use model router, you're only billed for the use of the underlying models as they're recruited to respond to prompts. The model routing function itself doesn't incur any extra charges.
5058

5159
You can monitor the costs of your model router deployment in the Azure portal.
5260

articles/ai-services/openai/concepts/models.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -56,13 +56,13 @@ A model that intelligently selects from a set of underlying chat models to respo
5656

5757
| Model | Region |
5858
|---|---|
59-
| `model-router` (2025-04-15) | East US 2 (Global Standard), Sweden Central (Global Standard)|
59+
| `model-router` (2025-05-19) | East US 2 (Global Standard), Sweden Central (Global Standard)|
6060

6161
### Capabilities
6262

6363
| Model ID | Description | Context Window | Max Output Tokens | Training Data (up to) |
6464
| --- | :--- |:--- |:---|:---: |
65-
| `model-router` (2025-04-15) | A model that intelligently selects from a set of underlying chat models to respond to a given prompt. | 200,000* | 32768 (GPT 4.1 series)</br> 100 K (o4-mini) | May 31, 2024 |
65+
| `model-router` (2025-05-19) | A model that intelligently selects from a set of underlying chat models to respond to a given prompt. | 200,000* | 32768 (GPT 4.1 series)</br> 100 K (o4-mini) | May 31, 2024 |
6666

6767
*Larger context windows are compatible with _some_ of the underlying models, which means an API call with a larger context will succeed only if the prompt happens to be routed to the right model, otherwise the call will fail.
6868

articles/ai-services/openai/how-to/model-router.md

Lines changed: 7 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,5 @@
11
---
2-
title: How to use model router (preview) in Azure OpenAI in Azure AI Foundry Models
3-
titleSuffix: Azure OpenAI
2+
title: How to use model router for Azure AI Foundry (preview)
43
description: Learn how to use the model router in Azure OpenAI to select the best model for your task.
54
author: PatrickFarley
65
ms.author: pafarley
@@ -11,20 +10,20 @@ ms.date: 04/17/2025
1110
manager: nitinme
1211
---
1312

14-
# Use Azure OpenAI model router (preview)
13+
# Use model router for Azure AI Foundry (preview)
1514

16-
Azure OpenAI model router is a deployable AI chat model that is trained to select the best large language model (LLM) to respond to a given prompt in real time. It uses a combination of preexisting models to provide high performance while saving on compute costs where possible. For more information on how model router works and its advantages and limitations, see the [Model router concepts guide](../concepts/model-router.md).
15+
Model router for Azure AI Foundry is a deployable AI chat model that is trained to select the best large language model (LLM) to respond to a given prompt in real time. It uses a combination of preexisting models to provide high performance while saving on compute costs where possible, all packaged as a single model deployment. For more information on how model router works and its advantages and limitations, see the [Model router concepts guide](../concepts/model-router.md).
1716

18-
You can access model router through the Completions API just as you would use a single base model like GPT-4.
17+
You can access model router through the Completions API just as you would use a single base model like GPT-4. The steps are the same as in the [Chat completions guide](/azure/ai-services/openai/how-to/chatgpt).
1918

2019
## Deploy a model router model
2120

22-
Model router is packaged as a single OpenAI model that you deploy. Follow the steps in the [resource deployment guide](/azure/ai-services/openai/how-to/create-resource), and in the **Create new deployment** step, find `Azure OpenAI model router` in the **Model** list. Select it, and then complete the rest of the deployment steps.
21+
Model router is packaged as a single Azure AI Foundry model that you deploy. Follow the steps in the [resource deployment guide](/azure/ai-services/openai/how-to/create-resource). In the **Create new deployment** step, find `model-router` in the **Models** list. Select it, and then complete the rest of the deployment steps.
2322

2423
> [!NOTE]
2524
> Consider that your deployment settings apply to all underlying chat models that model router uses.
2625
> - You don't need to deploy the underlying chat models separately. Model router works independently of your other deployed models.
27-
> - You select a content filter when you deploy the model router model (or you can apply a filter later). The content filter is applied to all activity to and from the model router: you don't set content filters for each of the underlying chat models.
26+
> - You select a content filter when you deploy the model router model (or you can apply a filter later). The content filter is applied to all content passed to and from the model router: you don't set content filters for each of the underlying chat models.
2827
> - Your tokens-per-minute rate limit setting is applied to all activity to and from the model router: you don't set rate limits for each of the underlying chat models.
2928
3029
## Use model router in chats
@@ -44,7 +43,7 @@ In the [Azure AI Foundry portal](https://ai.azure.com/), you can navigate to you
4443
4544
### Output format
4645

47-
The JSON response you receive from a model router model looks like the following.
46+
The JSON response you receive from a model router model is identical to the standard chat completions API response. Note that the `"model"` field reveals which underlying model was selected to respond to the prompt.
4847

4948
```json
5049
{

articles/ai-services/openai/quotas-limits.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -94,8 +94,8 @@ The following sections provide you with a quick guide to the default quotas and
9494

9595
| Model|Tier| Quota Limit in tokens per minute (TPM) | Requests per minute |
9696
|---|---|:---:|:---:|
97-
| `model-router` (2025-04-15) | Enterprise Tier | 10 M | 10 K |
98-
| `model-router` (2025-04-15) | Default | 1 M | 1 K |
97+
| `model-router` (2025-05-19) | Enterprise Tier | 10 M | 10 K |
98+
| `model-router` (2025-05-19) | Default | 1 M | 1 K |
9999

100100

101101
## computer-use-preview global standard rate limits

articles/ai-services/openai/whats-new.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@ Spotlighting is a sub-feature of prompt shields that enhances protection against
3131

3232
### Model router (preview)
3333

34-
Azure OpenAI model router is a deployable AI chat model that automatically selects the best underlying chat model to respond to a given prompt. For more information on how model router works and its advantages and limitations, see the [Model router concepts guide](./concepts/model-router.md). To use model router with the Completions API, follow the [How-to guide](./concepts/model-router.md).
34+
Model router for Azure AI Foundry is a deployable AI chat model that automatically selects the best underlying chat model to respond to a given prompt. For more information on how model router works and its advantages and limitations, see the [Model router concepts guide](./concepts/model-router.md). To use model router with the Completions API, follow the [How-to guide](./concepts/model-router.md).
3535

3636
## April 2025
3737

0 commit comments

Comments
 (0)