Skip to content

Commit 297866f

Browse files
authored
Merge pull request #2526 from MicrosoftDocs/main
1/27/2025 AM Publish
2 parents b861070 + 752e722 commit 297866f

11 files changed

+45
-41
lines changed

articles/ai-foundry/model-inference/includes/code-create-chat-client-entra.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -23,11 +23,11 @@ Then, you can use the package to consume the model. The following example shows
2323
```python
2424
import os
2525
from azure.ai.inference import ChatCompletionsClient
26-
from azure.identity import AzureDefaultCredential
26+
from azure.identity import DefaultAzureCredential
2727

2828
model = ChatCompletionsClient(
2929
endpoint="https://<resource>.services.ai.azure.com/models",
30-
credential=AzureDefaultCredential(),
30+
credential=DefaultAzureCredential(),
3131
model="mistral-large-2407",
3232
)
3333
```
@@ -45,11 +45,11 @@ Then, you can use the package to consume the model. The following example shows
4545
```javascript
4646
import ModelClient from "@azure-rest/ai-inference";
4747
import { isUnexpected } from "@azure-rest/ai-inference";
48-
import { AzureDefaultCredential } from "@azure/identity";
48+
import { DefaultAzureCredential } from "@azure/identity";
4949

5050
const client = new ModelClient(
5151
"https://<resource>.services.ai.azure.com/models",
52-
new AzureDefaultCredential(),
52+
new DefaultAzureCredential(),
5353
"mistral-large-2407"
5454
);
5555
```
@@ -81,7 +81,7 @@ Then, you can use the package to consume the model. The following example shows
8181
```csharp
8282
ChatCompletionsClient client = new ChatCompletionsClient(
8383
new Uri("https://<resource>.services.ai.azure.com/models"),
84-
new AzureDefaultCredential(includeInteractiveCredentials: true),
84+
new DefaultAzureCredential(includeInteractiveCredentials: true),
8585
"mistral-large-2407"
8686
);
8787
```

articles/ai-foundry/model-inference/includes/code-create-embeddings-client.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -36,11 +36,11 @@ If you are using an endpoint with support for Entra ID, you can create your clie
3636
```python
3737
import os
3838
from azure.ai.inference import EmbeddingsClient
39-
from azure.identity import AzureDefaultCredential
39+
from azure.identity import DefaultAzureCredential
4040

4141
client = EmbeddingsClient(
4242
endpoint="https://<resource>.services.ai.azure.com/models",
43-
credential=AzureDefaultCredential(),
43+
credential=DefaultAzureCredential(),
4444
)
4545
```
4646

@@ -72,11 +72,11 @@ For endpoint with support for Microsoft Entra ID, you can create your client as
7272
```javascript
7373
import ModelClient from "@azure-rest/ai-inference";
7474
import { isUnexpected } from "@azure-rest/ai-inference";
75-
import { AzureDefaultCredential } from "@azure/identity";
75+
import { DefaultAzureCredential } from "@azure/identity";
7676

7777
const client = new ModelClient(
7878
"https://<resource>.services.ai.azure.com/models",
79-
new AzureDefaultCredential()
79+
new DefaultAzureCredential()
8080
);
8181
```
8282

articles/ai-services/openai/how-to/prompt-caching.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -75,16 +75,16 @@ A single character difference in the first 1,024 tokens will result in a cache m
7575

7676
## What is cached?
7777

78-
The o1-series models are text only and don't support system messages, images, tool use/function calling, or structured outputs. This limits the efficacy of prompt caching for these models to the user/assistant portions of the messages array which are less likely to have an identical 1024 token prefix.
78+
o1-series models feature support varies by model. For more details, see our dedicated [reasoning models guide](./reasoning.md).
7979

8080
Prompt caching is supported for:
8181

8282
|**Caching supported**|**Description**|**Supported models**|
8383
|--------|--------|--------|
84-
| **Messages** | The complete messages array: system, user, and assistant content | `gpt-4o`<br/>`gpt-4o-mini`<br/>`gpt-4o-realtime-preview` (version 2024-12-17) |
85-
| **Images** | Images included in user messages, both as links or as base64-encoded data. The detail parameter must be set the same across requests. | `gpt-4o`<br/>`gpt-4o-mini` |
86-
| **Tool use** | Both the messages array and tool definitions. | `gpt-4o`<br/>`gpt-4o-mini`<br/>`gpt-4o-realtime-preview` (version 2024-12-17) |
87-
| **Structured outputs** | Structured output schema is appended as a prefix to the system message. | `gpt-4o`<br/>`gpt-4o-mini` |
84+
| **Messages** | The complete messages array: system, developer, user, and assistant content | `gpt-4o`<br/>`gpt-4o-mini`<br/>`gpt-4o-realtime-preview` (version 2024-12-17) <br> `o1` (version 2024-12-17) |
85+
| **Images** | Images included in user messages, both as links or as base64-encoded data. The detail parameter must be set the same across requests. | `gpt-4o`<br/>`gpt-4o-mini` <br> `o1` (version 2024-12-17) |
86+
| **Tool use** | Both the messages array and tool definitions. | `gpt-4o`<br/>`gpt-4o-mini`<br/>`gpt-4o-realtime-preview` (version 2024-12-17) <br> `o1` (version 2024-12-17) |
87+
| **Structured outputs** | Structured output schema is appended as a prefix to the system message. | `gpt-4o`<br/>`gpt-4o-mini` <br> `o1` (version 2024-12-17) |
8888

8989
To improve the likelihood of cache hits occurring, you should structure your requests such that repetitive content occurs at the beginning of the messages array.
9090

articles/ai-studio/how-to/create-azure-ai-resource.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -111,7 +111,7 @@ For hubs that use CMK encryption mode, you can update the encryption key to a ne
111111

112112
### Update Azure Application Insights and Azure Container Registry
113113

114-
To use custom environments for Prompt Flow, you're required to configure an Azure Container Registry for your hub. To use Azure Application Insights for Prompt Flow deployments, a configured Azure Application Insights resource is required for your hub. Updating the workspace-attached Azure Container Registry or ApplicationInsights resources may break lineage of previous jobs, deployed inference endpoints, or your ability to rerun earlier jobs in the workspace.
114+
To use custom environments for Prompt Flow, you're required to configure an Azure Container Registry for your hub. To use Azure Application Insights for Prompt Flow deployments, a configured Azure Application Insights resource is required for your hub. Updating the workspace-attached Azure Container Registry or Application Insights resources may break lineage of previous jobs, deployed inference endpoints, or your ability to rerun earlier jobs in the workspace. After association with an Azure AI Foundry hub, Azure Container Registry and Application Insights resources cannot be disassociated (set to null).
115115

116116
You can use the Azure portal, Azure SDK/CLI options, or the infrastructure-as-code templates to update both Azure Application Insights and Azure Container Registry for the hub.
117117

articles/ai-studio/reference/reference-model-inference-api.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -115,11 +115,11 @@ If you are using an endpoint with support for Entra ID, you can create your clie
115115
```python
116116
import os
117117
from azure.ai.inference import ChatCompletionsClient
118-
from azure.identity import AzureDefaultCredential
118+
from azure.identity import DefaultAzureCredential
119119

120120
model = ChatCompletionsClient(
121121
endpoint=os.environ["AZUREAI_ENDPOINT_URL"],
122-
credential=AzureDefaultCredential(),
122+
credential=DefaultAzureCredential(),
123123
)
124124
```
125125

@@ -151,11 +151,11 @@ For endpoint with support for Microsoft Entra ID, you can create your client as
151151
```javascript
152152
import ModelClient from "@azure-rest/ai-inference";
153153
import { isUnexpected } from "@azure-rest/ai-inference";
154-
import { AzureDefaultCredential } from "@azure/identity";
154+
import { DefaultAzureCredential } from "@azure/identity";
155155

156156
const client = new ModelClient(
157157
process.env.AZUREAI_ENDPOINT_URL,
158-
new AzureDefaultCredential()
158+
new DefaultAzureCredential()
159159
);
160160
```
161161

articles/machine-learning/reference-model-inference-api.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -108,11 +108,11 @@ If you are using an endpoint with support for Entra ID, you can create your clie
108108
```python
109109
import os
110110
from azure.ai.inference import ChatCompletionsClient
111-
from azure.identity import AzureDefaultCredential
111+
from azure.identity import DefaultAzureCredential
112112

113113
client = ChatCompletionsClient(
114114
endpoint=os.environ["AZUREAI_ENDPOINT_URL"],
115-
credential=AzureDefaultCredential(),
115+
credential=DefaultAzureCredential(),
116116
)
117117
```
118118

@@ -144,11 +144,11 @@ For endpoint with support for Microsoft Entra ID, you can create your client as
144144
```javascript
145145
import ModelClient from "@azure-rest/ai-inference";
146146
import { isUnexpected } from "@azure-rest/ai-inference";
147-
import { AzureDefaultCredential } from "@azure/identity";
147+
import { DefaultAzureCredential } from "@azure/identity";
148148

149149
const client = new ModelClient(
150150
process.env.AZUREAI_ENDPOINT_URL,
151-
new AzureDefaultCredential()
151+
new DefaultAzureCredential()
152152
);
153153
```
154154

articles/search/monitor-azure-cognitive-search-data-reference.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
---
22
title: Monitoring data reference for Azure AI Search
33
description: This article contains important reference material you need when you monitor Azure AI Search.
4-
ms.date: 02/15/2024
4+
ms.date: 01/27/2025
55
ms.custom: horz-monitor
66
ms.topic: reference
77
author: HeidiSteen

articles/search/monitor-azure-cognitive-search.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
---
22
title: Monitor Azure AI Search
33
description: Start here to learn how to monitor Azure AI Search.
4-
ms.date: 02/15/2024
4+
ms.date: 01/27/2025
55
ms.custom: horz-monitor
66
ms.topic: conceptual
77
author: HeidiSteen

articles/search/search-monitor-queries.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ ms.service: azure-ai-search
1010
ms.custom:
1111
- ignite-2023
1212
ms.topic: conceptual
13-
ms.date: 02/21/2024
13+
ms.date: 01/27/2025
1414
---
1515

1616
# Monitor query requests in Azure AI Search

articles/search/search-region-support.md

Lines changed: 12 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ ms.author: heidist
99
ms.service: azure-ai-search
1010
ms.topic: conceptual
1111
ms.custom: references_regions
12-
ms.date: 11/19/2024
12+
ms.date: 01/27/2025
1313

1414
---
1515

@@ -19,15 +19,17 @@ This article identifies the cloud regions in which Azure AI Search is available.
1919

2020
## Features subject to regional availability
2121

22-
| Feature | Availability |
23-
|---------|--------------|
24-
| [Extra capacity](search-limits-quotas-capacity.md#service-limits) | Higher capacity partitions became available in selected regions starting in April 2024 with a second wave following in May 2024. If you're using an older search service, create a new search service to benefit from more capacity at the same billing rate. To check existing capacity, [find your search service](https://portal.azure.com/#blade/HubsExtension/BrowseResourceBlade/resourceType/Microsoft.Search%2FsearchServices) and select the **Properties** tab in the middle of the Overview page. To check search service age, follow [these instructions](vector-search-index-size.md#how-to-check-service-creation-date). Currently, there are just a few regions that *don't* offer higher capacity partitions. Regional support for extra capacity is noted in the footnotes of this article.|
25-
| [Availability zones](search-reliability.md#availability-zone-support) | Divides a region's data centers into distinct physical location groups, providing high-availability within the same geo. Regional support is noted in this article. |
26-
| [AI service integration](cognitive-search-concept-intro.md) | Refers to skills that make internal calls to Azure AI for enrichment and transformation during indexing. Integration requires that Azure AI Search coexists with an [Azure AI multi-service account](/azure/ai-services/multi-service-resource) in the same physical region. Regional support is noted in this article. |
27-
| [Azure OpenAI integration](vector-search-integrated-vectorization.md) | Refers to skills and vectorizers that make internal calls to deployed embedding and chat models on Azure OpenAI. Check [Azure OpenAI model region availability](/azure/ai-services/openai/concepts/models#model-summary-table-and-region-availability) for the most current list of regions for each embedding and chat model. Specific Azure OpenAI models are in fewer regions, so be sure to check for joint regional availability before installing.|
28-
| [Azure AI Foundry integration](vector-search-integrated-vectorization-ai-studio.md) | Refers to skills and vectorizers that make internal calls to the models hosted in the model catalog. Check [Azure AI Foundry region availability](/azure/ai-studio/reference/region-support) for the most current list of regions. |
29-
| [Azure AI Vision 4.0 multimodal APIs for image vectorization](search-get-started-portal-image-search.md) | Refers to skills and vectorizers that call the multimodal embedding API. Check the [Azure AI Vision region list](/azure/ai-services/computer-vision/overview-image-analysis#region-availability) for joint regional availability. |
30-
| [Semantic ranker](semantic-search-overview.md) | Takes a dependency on Microsoft-hosted models in specific regions. Regional support is noted in this article. |
22+
Some features take a dependency on other Azure services or infrastructure that are subject to regional availability. If you need a specific feature, make sure it's available in the desired region.
23+
24+
| Feature | Description | Availability |
25+
|---------|-------------|--------------|
26+
| [Extra capacity](search-limits-quotas-capacity.md#service-limits) | Higher capacity partitions became available in selected regions starting in April 2024 with a second wave following in May 2024. Currently, there are just a few regions that *don't* offer higher capacity partitions. If you're using an older search service, create a new search service to benefit from more capacity at the same billing rate. | Regional support for extra capacity is noted in the footnotes of this article. <p>Check [service age](vector-search-index-size.md#how-to-check-service-creation-date) to see if your search service was created after high capacity partitions became available. <p>To check the capacity of an existing service, [find your search service](https://portal.azure.com/#blade/HubsExtension/BrowseResourceBlade/resourceType/Microsoft.Search%2FsearchServices) and select the **Properties** tab in the middle of the Overview page.|
27+
| [Availability zones](search-reliability.md#availability-zone-support) | Divides a region's data centers into distinct physical location groups, providing high-availability within the same geo. | Regional support is noted in this article. |
28+
| [Semantic ranker](semantic-search-overview.md) | Takes a dependency on Microsoft-hosted models in specific regions. | Regional support is noted in this article. |
29+
| [AI service integration](cognitive-search-concept-intro.md) | Refers to [built-in skills](cognitive-search-predefined-skills.md) that make internal calls to Azure AI for enrichment and transformation during indexing. Integration requires that Azure AI Search coexists with an [Azure AI multi-service account](/azure/ai-services/multi-service-resource) in the same physical region. You can bypass region requirements if you use [identity-based connections](cognitive-search-attach-cognitive-services.md#bill-through-a-keyless-connection), currently in public review. | Regional support is noted in this article. |
30+
| [Azure OpenAI integration](vector-search-integrated-vectorization.md) | Refers to the AzureOpenAIEmbedding skill and vectorizer that make internal calls to deployed embedding models on Azure OpenAI. | Check [Azure OpenAI model region availability](/azure/ai-services/openai/concepts/models#model-summary-table-and-region-availability) for the most current list of regions for each embedding and chat model. Specific Azure OpenAI models are in fewer regions, so check for model availability first, and then verify Azure AI Search is available in the same region.|
31+
| [Azure AI Foundry integration](vector-search-integrated-vectorization-ai-studio.md) | Refers to skills and vectorizers that make internal calls to the models hosted in the model catalog. | Check [Azure AI Foundry region availability](/azure/ai-studio/reference/region-support) for the most current list of regions. |
32+
| [Azure AI Vision 4.0 multimodal APIs](search-get-started-portal-image-search.md) | Refers to the Azure AI Vision multimodal embeddings skill and vectorizer that call the multimodal embedding API. | Check the [Azure AI Vision region list](/azure/ai-services/computer-vision/overview-image-analysis#region-availability) first, and then verify Azure AI Search is available in the same region.|
3133

3234
## Azure Public regions
3335

0 commit comments

Comments
 (0)