You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-services/agents/concepts/model-region-support.md
+49-22Lines changed: 49 additions & 22 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -22,46 +22,73 @@ Azure OpenAI provides customers with choices on the hosting structure that fits
22
22
-**Standard** is offered with a global deployment option, routing traffic globally to provide higher throughput.
23
23
-**Provisioned** is also offered with a global deployment option, allowing customers to purchase and deploy provisioned throughput units across Azure global infrastructure.
24
24
25
-
All deployments can perform the exact same inference operations, however the billing, scale, and performance are substantially different. To learn more about Azure OpenAI deployment types see our [deployment types guide](../../openai/how-to/deployment-types.md).
25
+
All deployments can perform the exact same inference operations, however the billing, scale, and performance are substantially different. To learn more about Azure OpenAI deployment types see [deployment types guide](../../openai/how-to/deployment-types.md).
26
26
27
27
Azure AI Agent Service supports the following Azure OpenAI models in the listed regions.
28
28
29
29
> [!NOTE]
30
30
> The following table is for pay-as-you-go. For information on Provisioned Throughput Unit (PTU) availability, see [provisioned throughput](../../openai/concepts/provisioned-throughput.md) in the Azure OpenAI documentation. `GlobalStandard` customers also have access to [global standard models](../../openai/concepts/models.md#global-standard-model-availability).
| swedencentral |||||| X | X | X | X | X | X | X | X | X |
45
+
| switzerlandnorth |||||||| X || X || X || X |
46
+
| uaenorth |||||||||||||||
47
+
| uksouth |||||||| X ||||| X | X |
48
+
| westus |||||| X | X | X | X || X || X ||
49
+
| westus3 |||||| X | X | X | X || X ||||
51
50
52
51
## Non-Microsoft models
53
52
54
53
The Azure AI Agent Service also supports the following models from the Azure AI Foundry model catalog.
55
54
56
55
* Meta-Llama-405B-Instruct
57
-
* Cohere-command-r-plus
58
-
* Cohere-command-r
59
56
60
57
To use these models, you can use [Azure AI Foundry portal](https://ai.azure.com/) to make a deployment, and then reference the deployment name in your agent. For example:
61
58
62
59
```python
63
60
agent = project_client.agents.create_agent( model="llama-3", name="my-agent", instructions="You are a helpful agent" )
64
61
```
62
+
## Azure AI Foundry models
63
+
64
+
### Models with tool-calling
65
+
66
+
To best support agentic scenarios, we recommend using models that support tool-calling. The Azure AI Agent Service currently supports all agent-compatible models from the Azure AI Foundry model catalog.
67
+
68
+
To use these models, use the [Azure AI Foundry portal](https://ai.azure.com/) to make a model deployment, then reference the deployment name in your agent. For example:
69
+
70
+
`agent = project_client.agents.create_agent( model="llama-3", name="my-agent", instructions="You are a helpful agent")`
71
+
72
+
> [!NOTE]
73
+
> This option should only be used for open-source models (for example, Cepstral, Mistral, Llama) and not for OpenAI models, which are natively supported in the service. This option should also only be used for models that support tool-calling.
74
+
75
+
### Models without tool-calling
76
+
77
+
Though tool-calling support is a core capability for agentic scenarios, we now provide the ability to use models that don’t support tool-calling in our API and SDK. This option can be helpful when you have specific use-cases that don’t require tool-calling.
78
+
79
+
The following steps will allow you to utilize any chat-completion model that is available through a [serverless API](/azure/ai-foundry/how-to/model-catalog-overview):
80
+
81
+
82
+
83
+
1. Deploy your desired model through serverless API. Model will show up on your **Models + Endpoints** page.
84
+
85
+
1. Click on model name to see model details, where you'll find your model's target URI and key.
86
+
87
+
1. Create a new Serverless connection on **Connected Resources** page, using the target URI and key.
88
+
89
+
The model can now be referenced in your code (`Target URI` + `@` + `Model Name`), for example:
Copy file name to clipboardExpand all lines: articles/ai-services/openai/api-version-deprecation.md
+12-3Lines changed: 12 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -18,17 +18,20 @@ This article is to help you understand the support lifecycle for the Azure OpenA
18
18
19
19
20
20
> [!NOTE]
21
-
> New API response objects may be added to the API response without version changes. We recommend you only parse the response objects you require.
21
+
> New API response objects may be added to the API response without version changes. We recommend you only parse the response objects you require.
22
+
>
23
+
> The latest Azure OpenAI spec uses OpenAPI 3.1. It is a known issue that this is currently not fully supported by [Azure API Management](/azure/api-management/api-management-key-concepts)
Copy file name to clipboardExpand all lines: articles/ai-services/openai/includes/api-surface.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -22,7 +22,7 @@ Each API surface/specification encapsulates a different set of Azure OpenAI capa
22
22
| API | Latest preview release | Latest GA release | Specifications | Description |
23
23
|:---|:----|:----|:----|:---|
24
24
| **Control plane** | [`2024-06-01-preview`](/rest/api/aiservices/accountmanagement/operation-groups?view=rest-aiservices-accountmanagement-2024-06-01-preview&preserve-view=true) | [`2024-10-01`](/rest/api/aiservices/accountmanagement/deployments/create-or-update?view=rest-aiservices-accountmanagement-2024-10-01&tabs=HTTP&preserve-view=true) | [Spec files](https://github.com/Azure/azure-rest-api-specs/tree/main/specification/cognitiveservices/resource-manager/Microsoft.CognitiveServices) | Azure OpenAI shares a common control plane with all other Azure AI Services. The control plane API is used for things like [creating Azure OpenAI resources](/rest/api/aiservices/accountmanagement/accounts/create?view=rest-aiservices-accountmanagement-2023-05-01&tabs=HTTP&preserve-view=true), [model deployment](/rest/api/aiservices/accountmanagement/deployments/create-or-update?view=rest-aiservices-accountmanagement-2023-05-01&tabs=HTTP&preserve-view=true), and other higher level resource management tasks. The control plane also governs what is possible to do with capabilities like Azure Resource Manager, Bicep, Terraform, and Azure CLI.|
25
-
| **Data plane - authoring** | `2025-04-01-preview` | `2024-10-21` | [Spec files](https://github.com/Azure/azure-rest-api-specs/tree/main/specification/cognitiveservices/data-plane/AzureOpenAI/authoring) | The data plane authoring API controls [fine-tuning](/rest/api/azureopenai/fine-tuning?view=rest-azureopenai-2024-08-01-preview&preserve-view=true), [file-upload](/rest/api/azureopenai/files/upload?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true), [ingestion jobs](/rest/api/azureopenai/ingestion-jobs/create?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true), [batch](/rest/api/azureopenai/batch?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true) and certain [model level queries](/rest/api/azureopenai/models/get?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true)
25
+
| **Data plane - authoring** | [`2025-04-01-preview`](/azure/ai-services/openai/authoring-reference-preview) | `2024-10-21` | [Spec files](https://github.com/Azure/azure-rest-api-specs/tree/main/specification/cognitiveservices/data-plane/AzureOpenAI/authoring) | The data plane authoring API controls [fine-tuning](/rest/api/azureopenai/fine-tuning?view=rest-azureopenai-2024-08-01-preview&preserve-view=true), [file-upload](/rest/api/azureopenai/files/upload?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true), [ingestion jobs](/rest/api/azureopenai/ingestion-jobs/create?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true), [batch](/rest/api/azureopenai/batch?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true) and certain [model level queries](/rest/api/azureopenai/models/get?view=rest-azureopenai-2024-08-01-preview&tabs=HTTP&preserve-view=true)
26
26
|**Data plane - inference**|[`2025-04-01-preview`](/azure/ai-services/openai/reference-preview#data-plane-inference)|[`2024-10-21`](/azure/ai-services/openai/reference#data-plane-inference)|[Spec files](https://github.com/Azure/azure-rest-api-specs/tree/main/specification/cognitiveservices/data-plane/AzureOpenAI/inference)| The data plane inference API provides the inference capabilities/endpoints for features like completions, chat completions, embeddings, audio, on your data, Dall-e, assistants, etc. |
0 commit comments