You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-services/openai/how-to/prompt-caching.md
+5-5Lines changed: 5 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -75,16 +75,16 @@ A single character difference in the first 1,024 tokens will result in a cache m
75
75
76
76
## What is cached?
77
77
78
-
The o1-series models are text only and don't support system messages, images, tool use/function calling, or structured outputs. This limits the efficacy of prompt caching for these models to the user/assistant portions of the messages array which are less likely to have an identical 1024 token prefix.
78
+
o1-series models feature support varies by model. For more details, see our dedicated [reasoning models guide](./reasoning.md).
|**Messages**| The complete messages array: system, user, and assistant content |`gpt-4o`<br/>`gpt-4o-mini`<br/>`gpt-4o-realtime-preview` (version 2024-12-17) |
85
-
|**Images**| Images included in user messages, both as links or as base64-encoded data. The detail parameter must be set the same across requests. |`gpt-4o`<br/>`gpt-4o-mini`|
86
-
|**Tool use**| Both the messages array and tool definitions. |`gpt-4o`<br/>`gpt-4o-mini`<br/>`gpt-4o-realtime-preview` (version 2024-12-17) |
87
-
|**Structured outputs**| Structured output schema is appended as a prefix to the system message. |`gpt-4o`<br/>`gpt-4o-mini`|
84
+
|**Messages**| The complete messages array: system, developer, user, and assistant content |`gpt-4o`<br/>`gpt-4o-mini`<br/>`gpt-4o-realtime-preview` (version 2024-12-17) <br> `o1` (version 2024-12-17) |
85
+
|**Images**| Images included in user messages, both as links or as base64-encoded data. The detail parameter must be set the same across requests. |`gpt-4o`<br/>`gpt-4o-mini`<br> `o1` (version 2024-12-17) |
86
+
|**Tool use**| Both the messages array and tool definitions. |`gpt-4o`<br/>`gpt-4o-mini`<br/>`gpt-4o-realtime-preview` (version 2024-12-17) <br> `o1` (version 2024-12-17) |
87
+
|**Structured outputs**| Structured output schema is appended as a prefix to the system message. |`gpt-4o`<br/>`gpt-4o-mini`<br> `o1` (version 2024-12-17) |
88
88
89
89
To improve the likelihood of cache hits occurring, you should structure your requests such that repetitive content occurs at the beginning of the messages array.
Copy file name to clipboardExpand all lines: articles/ai-studio/how-to/create-azure-ai-resource.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -111,7 +111,7 @@ For hubs that use CMK encryption mode, you can update the encryption key to a ne
111
111
112
112
### Update Azure Application Insights and Azure Container Registry
113
113
114
-
To use custom environments for Prompt Flow, you're required to configure an Azure Container Registry for your hub. To use Azure Application Insights for Prompt Flow deployments, a configured Azure Application Insights resource is required for your hub. Updating the workspace-attached Azure Container Registry or ApplicationInsights resources may break lineage of previous jobs, deployed inference endpoints, or your ability to rerun earlier jobs in the workspace.
114
+
To use custom environments for Prompt Flow, you're required to configure an Azure Container Registry for your hub. To use Azure Application Insights for Prompt Flow deployments, a configured Azure Application Insights resource is required for your hub. Updating the workspace-attached Azure Container Registry or Application Insights resources may break lineage of previous jobs, deployed inference endpoints, or your ability to rerun earlier jobs in the workspace. After association with an Azure AI Foundry hub, Azure Container Registry and Application Insights resources cannot be disassociated (set to null).
115
115
116
116
You can use the Azure portal, Azure SDK/CLI options, or the infrastructure-as-code templates to update both Azure Application Insights and Azure Container Registry for the hub.
Copy file name to clipboardExpand all lines: articles/search/search-region-support.md
+12-10Lines changed: 12 additions & 10 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -9,7 +9,7 @@ ms.author: heidist
9
9
ms.service: azure-ai-search
10
10
ms.topic: conceptual
11
11
ms.custom: references_regions
12
-
ms.date: 11/19/2024
12
+
ms.date: 01/27/2025
13
13
14
14
---
15
15
@@ -19,15 +19,17 @@ This article identifies the cloud regions in which Azure AI Search is available.
19
19
20
20
## Features subject to regional availability
21
21
22
-
| Feature | Availability |
23
-
|---------|--------------|
24
-
|[Extra capacity](search-limits-quotas-capacity.md#service-limits)| Higher capacity partitions became available in selected regions starting in April 2024 with a second wave following in May 2024. If you're using an older search service, create a new search service to benefit from more capacity at the same billing rate. To check existing capacity, [find your search service](https://portal.azure.com/#blade/HubsExtension/BrowseResourceBlade/resourceType/Microsoft.Search%2FsearchServices) and select the **Properties** tab in the middle of the Overview page. To check search service age, follow [these instructions](vector-search-index-size.md#how-to-check-service-creation-date). Currently, there are just a few regions that *don't* offer higher capacity partitions. Regional support for extra capacity is noted in the footnotes of this article.|
25
-
|[Availability zones](search-reliability.md#availability-zone-support)| Divides a region's data centers into distinct physical location groups, providing high-availability within the same geo. Regional support is noted in this article. |
26
-
|[AI service integration](cognitive-search-concept-intro.md)| Refers to skills that make internal calls to Azure AI for enrichment and transformation during indexing. Integration requires that Azure AI Search coexists with an [Azure AI multi-service account](/azure/ai-services/multi-service-resource) in the same physical region. Regional support is noted in this article. |
27
-
|[Azure OpenAI integration](vector-search-integrated-vectorization.md)| Refers to skills and vectorizers that make internal calls to deployed embedding and chat models on Azure OpenAI. Check [Azure OpenAI model region availability](/azure/ai-services/openai/concepts/models#model-summary-table-and-region-availability) for the most current list of regions for each embedding and chat model. Specific Azure OpenAI models are in fewer regions, so be sure to check for joint regional availability before installing.|
28
-
|[Azure AI Foundry integration](vector-search-integrated-vectorization-ai-studio.md)| Refers to skills and vectorizers that make internal calls to the models hosted in the model catalog. Check [Azure AI Foundry region availability](/azure/ai-studio/reference/region-support) for the most current list of regions. |
29
-
|[Azure AI Vision 4.0 multimodal APIs for image vectorization](search-get-started-portal-image-search.md)| Refers to skills and vectorizers that call the multimodal embedding API. Check the [Azure AI Vision region list](/azure/ai-services/computer-vision/overview-image-analysis#region-availability) for joint regional availability. |
30
-
|[Semantic ranker](semantic-search-overview.md)| Takes a dependency on Microsoft-hosted models in specific regions. Regional support is noted in this article. |
22
+
Some features take a dependency on other Azure services or infrastructure that are subject to regional availability. If you need a specific feature, make sure it's available in the desired region.
23
+
24
+
| Feature | Description | Availability |
25
+
|---------|-------------|--------------|
26
+
|[Extra capacity](search-limits-quotas-capacity.md#service-limits)| Higher capacity partitions became available in selected regions starting in April 2024 with a second wave following in May 2024. Currently, there are just a few regions that *don't* offer higher capacity partitions. If you're using an older search service, create a new search service to benefit from more capacity at the same billing rate. | Regional support for extra capacity is noted in the footnotes of this article. <p>Check [service age](vector-search-index-size.md#how-to-check-service-creation-date) to see if your search service was created after high capacity partitions became available. <p>To check the capacity of an existing service, [find your search service](https://portal.azure.com/#blade/HubsExtension/BrowseResourceBlade/resourceType/Microsoft.Search%2FsearchServices) and select the **Properties** tab in the middle of the Overview page.|
27
+
|[Availability zones](search-reliability.md#availability-zone-support)| Divides a region's data centers into distinct physical location groups, providing high-availability within the same geo. | Regional support is noted in this article. |
28
+
|[Semantic ranker](semantic-search-overview.md)| Takes a dependency on Microsoft-hosted models in specific regions. | Regional support is noted in this article. |
29
+
|[AI service integration](cognitive-search-concept-intro.md)| Refers to [built-in skills](cognitive-search-predefined-skills.md) that make internal calls to Azure AI for enrichment and transformation during indexing. Integration requires that Azure AI Search coexists with an [Azure AI multi-service account](/azure/ai-services/multi-service-resource) in the same physical region. You can bypass region requirements if you use [identity-based connections](cognitive-search-attach-cognitive-services.md#bill-through-a-keyless-connection), currently in public review. | Regional support is noted in this article. |
30
+
|[Azure OpenAI integration](vector-search-integrated-vectorization.md)| Refers to the AzureOpenAIEmbedding skill and vectorizer that make internal calls to deployed embedding models on Azure OpenAI. | Check [Azure OpenAI model region availability](/azure/ai-services/openai/concepts/models#model-summary-table-and-region-availability) for the most current list of regions for each embedding and chat model. Specific Azure OpenAI models are in fewer regions, so check for model availability first, and then verify Azure AI Search is available in the same region.|
31
+
|[Azure AI Foundry integration](vector-search-integrated-vectorization-ai-studio.md)| Refers to skills and vectorizers that make internal calls to the models hosted in the model catalog. | Check [Azure AI Foundry region availability](/azure/ai-studio/reference/region-support) for the most current list of regions. |
32
+
|[Azure AI Vision 4.0 multimodal APIs](search-get-started-portal-image-search.md)| Refers to the Azure AI Vision multimodal embeddings skill and vectorizer that call the multimodal embedding API. | Check the [Azure AI Vision region list](/azure/ai-services/computer-vision/overview-image-analysis#region-availability) first, and then verify Azure AI Search is available in the same region.|
0 commit comments