Skip to content

Commit 2604c68

Browse files
Learn Build Service GitHub AppLearn Build Service GitHub App
authored andcommitted
Merging changes synced from https://github.com/MicrosoftDocs/azure-docs-pr (branch live)
2 parents 1e22486 + 1aba0e1 commit 2604c68

File tree

1,234 files changed

+13054
-5826
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

1,234 files changed

+13054
-5826
lines changed

.openpublishing.redirection.json

Lines changed: 26 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2394,6 +2394,11 @@
23942394
"redirect_url": "/azure/architecture/web-apps/guides/multi-region-app-service/multi-region-app-service",
23952395
"redirect_document_id": false
23962396
},
2397+
{
2398+
"source_path_from_root": "/articles/app-service/app-service-configure-premium-tier.md",
2399+
"redirect_url": "/azure/app-service/app-service-configure-premium-v4-tier",
2400+
"redirect_document_id": false
2401+
},
23972402
{
23982403
"source_path_from_root": "/articles/public-multi-access-edge-compute-mec/tutorial-create-vm-using-python-sdk.md",
23992404
"redirect_url": "/previous-versions/azure/public-multi-access-edge-compute-mec/tutorial-create-vm-using-python-sdk",
@@ -6851,7 +6856,27 @@
68516856
},
68526857
{
68536858
"source_path_from_root": "/articles/azure-functions/migration.md",
6854-
"redirect_url": "/azure/azure-functions/migration/migrate-lambda-workloads-overview",
6859+
"redirect_url": "/azure/azure-functions/migration/migrate-aws-lambda-to-azure-functions",
6860+
"redirect_document_id": false
6861+
},
6862+
{
6863+
"source_path": "articles/azure-functions/migration/migrate-lambda-workloads-overview.md",
6864+
"redirect_url": "/azure/azure-functions/migration/migrate-aws-lambda-to-azure-functions",
6865+
"redirect_document_id": false
6866+
},
6867+
{
6868+
"source_path": "articles/azure-functions/migration/discover-lambda-workloads.md",
6869+
"redirect_url": "/azure/azure-functions/migration/migrate-aws-lambda-to-azure-functions",
6870+
"redirect_document_id": false
6871+
},
6872+
{
6873+
"source_path": "articles/azure-functions/migration/assess-lambda-workloads.md",
6874+
"redirect_url": "/azure/azure-functions/migration/migrate-aws-lambda-to-azure-functions",
6875+
"redirect_document_id": false
6876+
},
6877+
{
6878+
"source_path": "articles/azure-functions/migration/migrate-lambda-workloads.md",
6879+
"redirect_url": "/azure/azure-functions/migration/migrate-aws-lambda-to-azure-functions",
68556880
"redirect_document_id": false
68566881
},
68576882
{
Lines changed: 43 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,43 @@
1+
---
2+
title: Include file
3+
description: Include file
4+
services: api-center
5+
author: dlepow
6+
7+
ms.service: azure-api-center
8+
ms.topic: include
9+
ms.date: 05/12/2025
10+
ms.author: danlep
11+
ms.custom: Include file
12+
---
13+
14+
## About MCP servers
15+
16+
AI agents are becoming widely adopted because of enhanced large language model (LLM) capabilities. However, even the most advanced models face limitations because of their isolation from external data. Each new data source potentially requires custom implementations to extract, prepare, and make data accessible for the models.
17+
18+
The [model context protocol](https://www.anthropic.com/news/model-context-protocol) (MCP) helps solve this problem. MCP is an open standard for connecting AI models and agents with external data sources such as local data sources (databases or computer files) or remote services (systems available over the internet, such as remote databases or APIs).
19+
20+
### MCP architecture
21+
22+
The following diagram illustrates the MCP architecture:
23+
24+
:::image type="content" source="media/about-mcp-servers/mcp-architecture.png" alt-text="Diagram of model context protocol (MCP) architecture.":::
25+
26+
The architecture consists of the following components:
27+
28+
| Component | Description |
29+
|----------------|-------------------------------------------------------------------------------------------------|
30+
| **MCP hosts** | LLM applications such as chat apps or AI assistants in your IDEs (like GitHub Copilot in Visual Studio Code) that need to access external capabilities |
31+
| **MCP clients**| Protocol clients, inside the host application, that maintain 1:1 connections with servers |
32+
| **MCP servers**| Lightweight programs that each expose specific capabilities and provide context, tools, and prompts to clients |
33+
| **MCP protocol**| Transport layer in the middle |
34+
35+
MCP follows a client-server architecture where a host application can connect to multiple servers. Whenever your MCP host or client needs a tool, it connects to the MCP server. The MCP server then connects to, for example, a database or an API. MCP hosts and servers connect with each other through the MCP protocol.
36+
37+
### Remote versus local MCP servers
38+
39+
MCP utilizes a client-host-server architecture built on [JSON-RPC 2.0 for messaging](https://modelcontextprotocol.io/docs/concepts/architecture). Communication between clients and servers occurs over defined transport layers, and supports primarily two modes of operation:
40+
41+
* **Remote MCP servers** - MCP clients connect to MCP servers over the internet, establishing a connection using HTTP and server-sent events (SSE), and authorizing the MCP client access to resources on the user's account using OAuth.
42+
43+
* **Local MCP servers** MCP clients connect to MCP servers on the same machine, using standard input/output as a local transport method.

articles/api-center/register-discover-mcp-server.md

Lines changed: 1 addition & 30 deletions
Original file line numberDiff line numberDiff line change
@@ -14,36 +14,7 @@ ms.author: danlep
1414

1515
This article describes how to use Azure API Center to maintain an inventory (or *registry*) of remote model context protocol (MCP) servers and help stakeholders discover them using the API Center portal. MCP servers expose backend APIs or data sources in a standard way to AI agents and models that consume them.
1616

17-
## About MCP servers
18-
19-
AI agents are becoming widely adopted because of enhanced large language model (LLM) capabilities. However, even the most advanced models face limitations because of their isolation from external data. Each new data source potentially requires custom implementations to extract, prepare, and make data accessible for the models.
20-
21-
The [model context protocol](https://www.anthropic.com/news/model-context-protocol) (MCP) helps solve this problem. MCP is an open standard for connecting AI models and agents with external data sources such as local data sources (databases or computer files) or remote services (systems available over the internet, such as remote databases or APIs).
22-
23-
### MCP architecture
24-
25-
The following diagram illustrates the MCP architecture:
26-
27-
:::image type="content" source="media/register-discover-mcp-server/mcp-architecture.png" alt-text="Diagram of model context protocol (MCP) architecture.":::
28-
29-
The architecture consists of the following components:
30-
31-
| Component | Description |
32-
|----------------|-------------------------------------------------------------------------------------------------|
33-
| **MCP hosts** | LLM applications such as chat apps or AI assistants in your IDEs (like GitHub Copilot in Visual Studio Code) that need to access external capabilities |
34-
| **MCP clients**| Protocol clients, inside the host application, that maintain 1:1 connections with servers |
35-
| **MCP servers**| Lightweight programs that each expose specific capabilities and provide context, tools, and prompts to clients |
36-
| **MCP protocol**| Transport layer in the middle |
37-
38-
MCP follows a client-server architecture where a host application can connect to multiple servers. Whenever your MCP host or client needs a tool, it connects to the MCP server. The MCP server then connects to, for example, a database or an API. MCP hosts and servers connect with each other through the MCP protocol.
39-
40-
### Remote versus local MCP servers
41-
42-
MCP utilizes a client-host-server architecture built on [JSON-RPC 2.0 for messaging](https://modelcontextprotocol.io/docs/concepts/architecture). Communication between clients and servers occurs over defined transport layers, and supports primarily two modes of operation:
43-
44-
* **Remote MCP servers** - MCP clients connect to MCP servers over the internet, establishing a connection using HTTP and server-sent events (SSE), and authorizing the MCP client access to resources on the user's account using OAuth.
45-
46-
* **Local MCP servers** MCP clients connect to MCP servers on the same machine, using standard input/output as a local transport method.
17+
[!INCLUDE [about-mcp-servers](includes/about-mcp-servers.md)]
4718

4819
## MCP servers in your API inventory
4920

articles/api-management/TOC.yml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -212,6 +212,8 @@
212212
href: azure-openai-api-from-specification.md
213213
- name: Authenticate and authorize to Azure OpenAI
214214
href: api-management-authenticate-authorize-azure-openai.md
215+
- name: Expose REST API as MCP server
216+
href: export-rest-mcp-server.md
215217
- name: Semantic caching for Azure OpenAI API requests
216218
href: azure-openai-enable-semantic-caching.md
217219
- name: Protect Azure OpenAI keys

articles/api-management/api-management-capacity.md

Lines changed: 39 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,12 @@
11
---
2-
title: Capacity metrics - Azure API Management | Microsoft Docs
2+
title: Capacity Metrics - Azure API Management
33
description: This article explains the capacity metrics in Azure API Management and how to make informed decisions about whether to scale an instance.
44
services: api-management
55
author: dlepow
66

77
ms.service: azure-api-management
88
ms.topic: how-to
9-
ms.date: 10/01/2024
9+
ms.date: 04/17/2025
1010
ms.author: danlep
1111
ms.custom: fasttrack-edit
1212
---
@@ -19,18 +19,18 @@ API Management provides [Azure Monitor metrics](api-management-howto-use-azure-m
1919

2020
This article explains the capacity metrics and how they behave, shows how to access capacity metrics in the Azure portal, and suggests when to consider scaling or upgrading your API Management instance.
2121

22-
[!INCLUDE [api-management-workspace-availability](../../includes/api-management-workspace-availability.md)]
23-
2422
> [!IMPORTANT]
25-
> This article introduces how to monitor and scale your Azure API Management instance based on capacity metrics. However, when an instance *reaches* its capacity, it won't throttle to prevent overload. Instead, it will act like an overloaded web server: increased latency, dropped connections, and timeout errors. API clients should be ready to handle these issues as they do with other external services, for example by using retry policies.
23+
> This article introduces how to monitor and scale your Azure API Management instance based on capacity metrics. However, when an instance *reaches* its capacity, it won't throttle to prevent overload. Instead, it will act like an overloaded web server: increased latency, dropped connections, and time-out errors. API clients should be ready to handle these issues as they do with other external services, for example by using retry policies.
24+
25+
[!INCLUDE [api-management-workspace-try-it](../../includes/api-management-workspace-try-it.md)]
2626

2727
## Prerequisites
2828

2929
To follow the steps in this article, you must have an API Management instance in one of the tiers that supports capacity metrics. For more information, see [Create an Azure API Management instance](get-started-create-service-instance.md).
3030

3131
## Available capacity metrics
3232

33-
Different capacity metrics are available in the [v2 service tiers](v2-service-tiers-overview.md) and classic tiers.
33+
Different capacity metrics are available in the [v2 service tiers](v2-service-tiers-overview.md), classic tiers, and [workspace gateways](workspaces-overview.md#workspace-gateway).
3434

3535
#### [v2 tiers](#tab/v2-tiers)
3636

@@ -66,6 +66,21 @@ Available aggregations for this metric are as follows.
6666

6767
[!INCLUDE [api-management-cpu-memory-capacity](../../includes/api-management-cpu-memory-capacity.md)]
6868

69+
#### [Workspace gateways](#tab/workspace-gateway)
70+
71+
In a workspace gateway, the following metrics are available:
72+
73+
* **CPU Utilization (%)** - The percentage of CPU capacity used by the gateway units.
74+
75+
* **Memory Utilization (%)** - The percentage of memory capacity used by the gateway units.
76+
77+
Available aggregations for these metrics are as follows.
78+
79+
* **Avg** - Average percentage of capacity used across gateway processes in every [unit](upgrade-and-scale.md) of a workspace gateway.
80+
* **Max** - Percentage of capacity in gateway process with the greatest consumption.
81+
82+
[!INCLUDE [api-management-cpu-memory-capacity](../../includes/api-management-cpu-memory-capacity.md)]
83+
6984
---
7085

7186
## Capacity metric behavior
@@ -85,7 +100,7 @@ The more complex operations on the requests are, the higher the capacity consump
85100

86101
Capacity metrics can also spike intermittently or be greater than zero even if no requests are being processed. It happens because of system- or platform-specific actions and should not be taken into consideration when deciding whether to scale an instance.
87102

88-
Although capacity metrics are designed to surface problems with your API Management instance, there are cases when problems won't be reflected in changes in these metrics. Additionally, low capacity metrics don't necessarily mean that your API Management instance isn't experiencing any problems.
103+
Although capacity metrics are designed to surface problems with your API Management instance (or workspace gateway), there are cases when problems won't be reflected in changes in these metrics. Additionally, low capacity metrics don't necessarily mean that your API Management instance isn't experiencing any problems.
89104

90105

91106
## Use the Azure portal to examine capacity metrics
@@ -128,19 +143,32 @@ Access metrics in the portal to understand how much capacity is used over time.
128143
> * You can use Azure Monitor [autoscaling](api-management-howto-autoscale.md) to automatically add an Azure API Management unit. Scaling operation can take around 30 minutes, so you should plan your rules accordingly.
129144
> * In multi-region deployments, only scaling the primary location is allowed.
130145
146+
#### [Workspace gateways](#tab/workspace-gateway)
147+
148+
1. Navigate to your API Management instance in the [Azure portal](https://portal.azure.com/).
149+
1. In the left menu, under **APIs**, select **Workspaces** > the workspace associated with the gateway that you want to monitor.
150+
1. In the left menu, under **Deployment + infrastructure**, select **Gateways** > the workspace gateway that you want to monitor.
151+
1. In the left menu, under **Monitoring**, select **Metrics**.
152+
1. Select the **CPU Utilization (%)** or **Memory Utilization (%)** metric from the available metrics. Choose the default **Avg** aggregation or select the **Max** aggregation to see the peak usage.
153+
1. Pick a desired timeframe from the top bar of the section.
154+
155+
> [!NOTE]
156+
> You can set a [metric alert](api-management-howto-use-azure-monitor.md#set-up-an-alert-rule) to let you know when something unexpected is happening. For example, get notifications when your workspace gateway has exceeded its expected peak CPU or Memory usage for more than 20 minutes.
157+
158+
131159
---
132160

133161
## Use capacity for scaling decisions
134162

135-
Use capacity metrics for making decisions whether to scale an API Management instance to accommodate more load. The following are general considerations:
163+
Use capacity metrics for making decisions whether to scale an API Management instance (or workspace gateway) to accommodate more load. The following are general considerations:
136164

137165
+ Look at a long-term trend and average.
138166
+ Ignore sudden spikes that are most likely not related to an increase in load (see [Capacity metric behavior](#capacity-metric-behavior) section for explanation).
139167
+ As a general rule, upgrade or scale your instance when a capacity metric value exceeds **60% - 70%** for a long period of time (for example, 30 minutes). Different values may work better for your service or scenario.
140-
+ If your instance is configured with only 1 unit, upgrade or scale your instance when a capacity metric value exceeds **40%** for a long period. This recommendation is based on the need to reserve capacity for guest OS updates in the underlying service platform.
168+
+ If your instance or workspace gateway is configured with only 1 unit, upgrade or scale it when a capacity metric value exceeds **40%** for a long period. This recommendation is based on the need to reserve capacity for guest OS updates in the underlying service platform.
141169

142-
>[!TIP]
143-
> If you are able to estimate your traffic beforehand, test your API Management instance on workloads you expect. You can increase the request load on your tenant gradually and monitor the value of the capacity metric that corresponds to your peak load. Follow the steps from the previous section to use Azure portal to understand how much capacity is used at any given time.
170+
> [!TIP]
171+
> If you are able to estimate your traffic beforehand, test your API Management instance or workspace gateway on workloads you expect. You can increase the request load gradually and monitor the value of the capacity metric that corresponds to your peak load. Follow the steps from the previous section to use Azure portal to understand how much capacity is used at any given time.
144172
145173
> [!IMPORTANT]
146174
> When defining autoscaling rules, make sure to avoid autoscaling flapping that will cause a scale in/scale out loop. To learn more, we recommend reading [Azure Monitor Autoscale's guidance around flapping](/azure/azure-monitor/autoscale/autoscale-flapping).

articles/api-management/api-management-faq.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -62,7 +62,7 @@ sections:
6262
* The [API Management REST API](/rest/api/apimanagement/)
6363
* Azure client SDKs for [.NET](/dotnet/api/overview/azure/api-management), [JavaScript](/javascript/api/overview/azure/api-management), [Java](/java/api/overview/azure/api-management), and other languages
6464
* [Azure PowerShell](/powershell/module/az.apimanagement/) and [Azure CLI](/cli/azure/apim)
65-
* [Azure Resource Manager](/azure/templates/microsoft.apimanagement/allversions) and Bicep templates. See example [quickstart templates](https://azure.microsoft.com/resources/templates/?Page=2&term=api+management).
65+
* [Azure Resource Manager](/azure/templates/microsoft.apimanagement/allversions) and Bicep files. See example [quickstart templates](https://azure.microsoft.com/resources/templates/?Page=2&term=api+management).
6666
* [Terraform](https://registry.terraform.io/providers/hashicorp/azurerm/latest/docs/resources/api_management)
6767
6868
- question: |

articles/api-management/api-management-features.md

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ author: dlepow
66

77
ms.service: azure-api-management
88
ms.topic: concept-article
9-
ms.date: 10/15/2024
9+
ms.date: 05/12/2025
1010
ms.author: danlep
1111
---
1212

@@ -48,6 +48,9 @@ Each API Management [pricing tier](api-management-key-concepts.md#api-management
4848
| Azure Monitor and Log Analytics request logs | No | Yes | Yes | Yes | Yes | Yes |Yes | Yes |
4949
| Application Insights request logs | Yes | Yes | Yes | Yes | Yes | Yes |Yes | Yes |
5050
| Static IP | No | Yes | Yes | No |Yes | No | Yes | No |
51+
| Export API to Power Platform | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes |
52+
| Export API to Postman | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes |
53+
| Export API to MCP server (preview) | No | No | Yes | No | Yes | No | Yes | No |
5154

5255
<sup>1</sup> Enables the use of Microsoft Entra ID (and Azure AD B2C) as an identity provider for user sign in on the developer portal.<br/>
5356
<sup>2</sup> Including related functionality such as users, groups, issues, applications, and email templates and notifications.<br/>

0 commit comments

Comments
 (0)