Skip to content

Commit 29e6e91

Browse files
authored
Merge branch 'MicrosoftDocs:main' into heidist-uuf
2 parents e155da2 + 82c8bad commit 29e6e91

File tree

12 files changed

+55
-60
lines changed

12 files changed

+55
-60
lines changed

articles/ai-foundry/model-inference/concepts/deployment-types.md

Lines changed: 17 additions & 25 deletions
Original file line numberDiff line numberDiff line change
@@ -13,43 +13,35 @@ ms.custom: ignite-2024, github-universe-2024
1313

1414
# Deployment types in Azure AI model inference
1515

16-
Azure AI model inference in Azure AI services provides customers with choices on the hosting structure that fits their business and usage patterns. The service offers two main types of deployment: **standard** and **provisioned**. Standard is offered with a global deployment option, routing traffic globally to provide higher throughput. Provisioned is also offered with a global deployment option, allowing customers to purchase and deploy provisioned throughput units across Azure global infrastructure.
16+
Azure AI model inference makes models available using the *model deployment* concept in Azure AI Services resources. *Model deployments* are also Azure resources and, when created, they give access to a given model under certain configurations. Such configuration includes the infrastructure require to process the requests.
1717

18-
All deployments can perform the exact same inference operations, however the billing, scale, and performance are substantially different. As part of your solution design, you need to make two key decisions:
18+
Azure AI model inference provides customers with choices on the hosting structure that fits their business and usage patterns. Those options are translated to different deployments types (or SKUs) that are available at model deployment time in the Azure AI Services resource.
1919

20-
- **Data residency needs**: global vs. regional resources
21-
- **Call volume**: standard vs. provisioned
20+
:::image type="content" source="../media/add-model-deployments/models-deploy-deployment-type.png" alt-text="Screenshot showing how to customize the deployment type for a given model deployment." lightbox="../media/add-model-deployments/models-deploy-deployment-type.png":::
2221

23-
Deployment types support varies by model and model provider. You can see which deployment type (SKU) each model supports in the [Models section](models.md).
22+
Different model providers offer different deployments SKUs that you can select from. When selecting a deployment type, consider your **data residency needs** and **call volume/capacity** requirements.
2423

25-
## Global versus regional deployment types
24+
## Deployment types for Azure OpenAI models
2625

27-
For standard and provisioned deployments, you have an option of two types of configurations within your resource – **global** or **regional**. Global standard is the recommended starting point.
26+
The service offers two main types of deployments: **standard** and **provisioned**. For a given deployment type, customers can align their workloads with their data processing requirements by choosing an Azure geography (`Standard` or `Provisioned-Managed`), Microsoft specified data zone (`DataZone-Standard` or `DataZone Provisioned-Managed`), or Global (`Global-Standard` or `Global Provisioned-Managed`) processing options.
2827

29-
Global deployments leverage Azure's global infrastructure, dynamically route customer traffic to the data center with best availability for the customer's inference requests. This means you get the highest initial throughput limits and best model availability with Global while still providing our uptime SLA and low latency. For high volume workloads above the specified usage tiers on standard and global standard, you may experience increased latency variation. For customers that require the lower latency variance at large workload usage, we recommend purchasing provisioned throughput.
28+
To learn more about deployment options for Azure OpenAI models see [Azure OpenAI documentation](../../../ai-services/openai/how-to/deployment-types.md).
3029

31-
Our global deployments are the first location for all new models and features. Customers with large throughput requirements should consider our provisioned deployment offering.
30+
## Deployment types for Models-as-a-Service models
3231

33-
## Standard
32+
Models from third-party model providers with pay-as-you-go billing (collectively called Models-as-a-Service), makes models available in Azure AI model inference under **standard** deployments with a Global processing option (`Global-Standard`).
3433

35-
Standard deployments provide a pay-per-call billing model on the chosen model. Provides the fastest way to get started as you only pay for what you consume. Models available in each region and throughput may be limited.
34+
Models-as-a-Service offers regional deployment options under [Serverless API endpoints](../../../ai-studio/how-to/deploy-models-serverless.md) in Azure AI Foundry. Prompts and outputs are processed within the geography specified during deployment. However, those deployments can't be accessed using the Azure AI model inference endpoint in Azure AI Services.
3635

37-
Standard deployments are optimized for low to medium volume workloads with high burstiness. Customers with high consistent volume may experience greater latency variability.
36+
### Global-Standard
3837

39-
Only Azure OpenAI models support this deployment type.
38+
Global deployments leverage Azure's global infrastructure to dynamically route traffic to the data center with best availability for each request. Global standard provides the highest default quota and eliminates the need to load balance across multiple resources. Data stored at rest remains in the designated Azure geography, while data may be processed for inferencing in any Azure location. Learn more about [data residency](https://azure.microsoft.com/explore/global-infrastructure/data-residency/).
4039

41-
## Global standard
40+
## Control deployment options
4241

43-
Global deployments are available in the same Azure AI services resources as non-global deployment types but allow you to leverage Azure's global infrastructure to dynamically route traffic to the data center with best availability for each request. Global standard provides the highest default quota and eliminates the need to load balance across multiple resources.
42+
Administrators can control which model deployment types are available to their users by using Azure Policies. Learn more about [How to control AI model deployment with custom policies](../../../ai-studio/how-to/custom-policy-model-deployment.md).
4443

45-
Customers with high consistent volume may experience greater latency variability. The threshold is set per model. For applications that require the lower latency variance at large workload usage, we recommend purchasing provisioned throughput if available.
44+
## Related content
4645

47-
## Global provisioned
48-
49-
Global deployments are available in the same Azure AI services resources as non-global deployment types but allow you to leverage Azure's global infrastructure to dynamically route traffic to the data center with best availability for each request. Global provisioned deployments provide reserved model processing capacity for high and predictable throughput using Azure global infrastructure.
50-
51-
Only Azure OpenAI models support this deployment type.
52-
53-
## Next steps
54-
55-
- [Quotas & limits](../quotas-limits.md)
46+
- [Quotas & limits](../quotas-limits.md)
47+
- [Data privacy, and security for Models-as-a-Service models](../../../ai-studio/how-to/concept-data-privacy.md)

articles/ai-foundry/model-inference/includes/code-create-chat-client-entra.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -23,11 +23,11 @@ Then, you can use the package to consume the model. The following example shows
2323
```python
2424
import os
2525
from azure.ai.inference import ChatCompletionsClient
26-
from azure.identity import AzureDefaultCredential
26+
from azure.identity import DefaultAzureCredential
2727

2828
model = ChatCompletionsClient(
2929
endpoint="https://<resource>.services.ai.azure.com/models",
30-
credential=AzureDefaultCredential(),
30+
credential=DefaultAzureCredential(),
3131
model="mistral-large-2407",
3232
)
3333
```
@@ -45,11 +45,11 @@ Then, you can use the package to consume the model. The following example shows
4545
```javascript
4646
import ModelClient from "@azure-rest/ai-inference";
4747
import { isUnexpected } from "@azure-rest/ai-inference";
48-
import { AzureDefaultCredential } from "@azure/identity";
48+
import { DefaultAzureCredential } from "@azure/identity";
4949

5050
const client = new ModelClient(
5151
"https://<resource>.services.ai.azure.com/models",
52-
new AzureDefaultCredential(),
52+
new DefaultAzureCredential(),
5353
"mistral-large-2407"
5454
);
5555
```
@@ -81,7 +81,7 @@ Then, you can use the package to consume the model. The following example shows
8181
```csharp
8282
ChatCompletionsClient client = new ChatCompletionsClient(
8383
new Uri("https://<resource>.services.ai.azure.com/models"),
84-
new AzureDefaultCredential(includeInteractiveCredentials: true),
84+
new DefaultAzureCredential(includeInteractiveCredentials: true),
8585
"mistral-large-2407"
8686
);
8787
```

articles/ai-foundry/model-inference/includes/code-create-embeddings-client.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -36,11 +36,11 @@ If you are using an endpoint with support for Entra ID, you can create your clie
3636
```python
3737
import os
3838
from azure.ai.inference import EmbeddingsClient
39-
from azure.identity import AzureDefaultCredential
39+
from azure.identity import DefaultAzureCredential
4040

4141
client = EmbeddingsClient(
4242
endpoint="https://<resource>.services.ai.azure.com/models",
43-
credential=AzureDefaultCredential(),
43+
credential=DefaultAzureCredential(),
4444
)
4545
```
4646

@@ -72,11 +72,11 @@ For endpoint with support for Microsoft Entra ID, you can create your client as
7272
```javascript
7373
import ModelClient from "@azure-rest/ai-inference";
7474
import { isUnexpected } from "@azure-rest/ai-inference";
75-
import { AzureDefaultCredential } from "@azure/identity";
75+
import { DefaultAzureCredential } from "@azure/identity";
7676

7777
const client = new ModelClient(
7878
"https://<resource>.services.ai.azure.com/models",
79-
new AzureDefaultCredential()
79+
new DefaultAzureCredential()
8080
);
8181
```
8282

articles/ai-foundry/model-inference/includes/configure-project-connection/portal.md

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,11 @@ zone_pivot_groups: azure-ai-models-deployment
1010

1111
[!INCLUDE [Header](intro.md)]
1212

13-
* An AI project connected to your Azure AI Services resource. You call follow the steps at [Configure Azure AI model inference service in my project](../../how-to/configure-project-connection.md) in Azure AI Foundry.
13+
* An AI project resource.
14+
15+
* The feature **Deploy models to Azure AI model inference service** on.
16+
17+
:::image type="content" source="../../media/quickstart-ai-project/ai-project-inference-endpoint.gif" alt-text="An animation showing how to turn on the Deploy models to Azure AI model inference service feature in Azure AI Foundry portal." lightbox="../../media/quickstart-ai-project/ai-project-inference-endpoint.gif":::
1418

1519
## Add a connection
1620

@@ -50,4 +54,4 @@ You can see the model deployments available in the connected resource by followi
5054

5155
5. The details page shows information about the specific deployment. If you want to test the model, you can use the option **Open in playground**.
5256

53-
6. The Azure AI Foundry playground is displayed, where you can interact with the given model.
57+
6. The Azure AI Foundry playground is displayed, where you can interact with the given model.

articles/ai-services/language-service/custom-named-entity-recognition/includes/use-pre-existing-resource.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -51,7 +51,7 @@ Make sure to enable **Custom text classification / Custom Named Entity Recogniti
5151
5. Select **Apply**.
5252

5353
>[!Important]
54-
> * Make sure that your **Language resource** has **storage blob data contributor** role assigned on the storage account you are connecting.
54+
> Make sure that the user making changes has **storage blob data contributor** role assigned for them.
5555
5656
### Add required roles
5757

articles/ai-services/openai/how-to/prompt-caching.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -75,16 +75,16 @@ A single character difference in the first 1,024 tokens will result in a cache m
7575

7676
## What is cached?
7777

78-
The o1-series models are text only and don't support system messages, images, tool use/function calling, or structured outputs. This limits the efficacy of prompt caching for these models to the user/assistant portions of the messages array which are less likely to have an identical 1024 token prefix.
78+
o1-series models feature support varies by model. For more details, see our dedicated [reasoning models guide](./reasoning.md).
7979

8080
Prompt caching is supported for:
8181

8282
|**Caching supported**|**Description**|**Supported models**|
8383
|--------|--------|--------|
84-
| **Messages** | The complete messages array: system, user, and assistant content | `gpt-4o`<br/>`gpt-4o-mini`<br/>`gpt-4o-realtime-preview` (version 2024-12-17) |
85-
| **Images** | Images included in user messages, both as links or as base64-encoded data. The detail parameter must be set the same across requests. | `gpt-4o`<br/>`gpt-4o-mini` |
86-
| **Tool use** | Both the messages array and tool definitions. | `gpt-4o`<br/>`gpt-4o-mini`<br/>`gpt-4o-realtime-preview` (version 2024-12-17) |
87-
| **Structured outputs** | Structured output schema is appended as a prefix to the system message. | `gpt-4o`<br/>`gpt-4o-mini` |
84+
| **Messages** | The complete messages array: system, developer, user, and assistant content | `gpt-4o`<br/>`gpt-4o-mini`<br/>`gpt-4o-realtime-preview` (version 2024-12-17) <br> `o1` (version 2024-12-17) |
85+
| **Images** | Images included in user messages, both as links or as base64-encoded data. The detail parameter must be set the same across requests. | `gpt-4o`<br/>`gpt-4o-mini` <br> `o1` (version 2024-12-17) |
86+
| **Tool use** | Both the messages array and tool definitions. | `gpt-4o`<br/>`gpt-4o-mini`<br/>`gpt-4o-realtime-preview` (version 2024-12-17) <br> `o1` (version 2024-12-17) |
87+
| **Structured outputs** | Structured output schema is appended as a prefix to the system message. | `gpt-4o`<br/>`gpt-4o-mini` <br> `o1` (version 2024-12-17) |
8888

8989
To improve the likelihood of cache hits occurring, you should structure your requests such that repetitive content occurs at the beginning of the messages array.
9090

articles/ai-services/translator/firewalls.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,19 +1,19 @@
11
---
2-
title: Translate behind firewalls - Translator
2+
title: Use Azure AI Translator to translate behind firewalls.
33
titleSuffix: Azure AI services
44
description: Azure AI Translator can translate behind firewalls using either domain-name or IP filtering.
55
#services: cognitive-services
66
author: laujan
77
manager: nitinme
88
ms.service: azure-ai-translator
99
ms.topic: conceptual
10-
ms.date: 07/09/2024
10+
ms.date: 01/27/2025
1111
ms.author: lajanuar
1212
---
1313

14-
# Use Translator behind firewalls
14+
# Use Azure AI Translator behind firewalls
1515

16-
Translator can translate behind firewalls using either [Domain-name](/azure/firewall/dns-settings#dns-proxy-configuration) or [IP filtering](#configure-firewall). Domain-name filtering is the preferred method.
16+
Azure AI Translator can translate behind firewalls using either [Domain-name](/azure/firewall/dns-settings#dns-proxy-configuration) or [IP filtering](#configure-firewall). Domain-name filtering is the preferred method.
1717

1818
If you still require IP filtering, you can get the [IP addresses details using service tag](/azure/virtual-network/service-tags-overview#discover-service-tags-by-using-downloadable-json-files). Translator is under the **CognitiveServicesManagement** service tag.
1919

articles/ai-services/translator/whats-new.md

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,13 +1,12 @@
11
---
22
title: What's new in Azure AI Translator?
33
titleSuffix: Azure AI services
4-
description: Learn of the latest changes to the Translator Service API.
4+
description: Learn about the latest changes to the Azure AI Translator Service API.
55
author: laujan
66
manager: nitinme
77
ms.service: azure-ai-translator
8-
ms.custom: build-2023
98
ms.topic: overview
10-
ms.date: 06/19/2024
9+
ms.date: 01/27/2025
1110
ms.author: lajanuar
1211
---
1312
<!-- markdownlint-disable MD024 -->
@@ -20,7 +19,7 @@ Bookmark this page to stay up to date with release notes, feature enhancements,
2019

2120
Translator is a language service that enables users to translate text and documents, helps entities expand their global outreach, and supports preservation of at-risk and endangered languages.
2221

23-
Translator service supports language translation for more than 100 languages. If your language community is interested in partnering with Microsoft to add your language to Translator, contact us via the [Translator community partner onboarding form](https://forms.office.com/pages/responsepage.aspx?id=v4j5cvGGr0GRqy180BHbR-riVR3Xj0tOnIRdZOALbM9UOU1aMlNaWFJOOE5YODhRR1FWVzY0QzU1OS4u).
22+
Azure AI Translator service supports language translation for more than 100 languages. If your language community is interested in partnering with Microsoft to add your language to Translator, contact us via the [Translator community partner onboarding form](https://forms.office.com/pages/responsepage.aspx?id=v4j5cvGGr0GRqy180BHbR-riVR3Xj0tOnIRdZOALbM9UOU1aMlNaWFJOOE5YODhRR1FWVzY0QzU1OS4u).
2423

2524
## May 2024
2625

articles/ai-studio/.openpublishing.redirection.ai-studio.json

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -207,7 +207,7 @@
207207
},
208208
{
209209
"source_path_from_root": "/articles/ai-studio/ai-services/how-to/content-safety.md",
210-
"redirect_url": "/azure/ai-foundry/model-inference/how-to/configure-content-safety",
210+
"redirect_url": "/azure/ai-foundry/model-inference/how-to/configure-content-filters",
211211
"redirect_document_id": false
212212
},
213213
{

articles/ai-studio/how-to/create-azure-ai-resource.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -111,7 +111,7 @@ For hubs that use CMK encryption mode, you can update the encryption key to a ne
111111

112112
### Update Azure Application Insights and Azure Container Registry
113113

114-
To use custom environments for Prompt Flow, you're required to configure an Azure Container Registry for your hub. To use Azure Application Insights for Prompt Flow deployments, a configured Azure Application Insights resource is required for your hub. Updating the workspace-attached Azure Container Registry or ApplicationInsights resources may break lineage of previous jobs, deployed inference endpoints, or your ability to rerun earlier jobs in the workspace.
114+
To use custom environments for Prompt Flow, you're required to configure an Azure Container Registry for your hub. To use Azure Application Insights for Prompt Flow deployments, a configured Azure Application Insights resource is required for your hub. Updating the workspace-attached Azure Container Registry or Application Insights resources may break lineage of previous jobs, deployed inference endpoints, or your ability to rerun earlier jobs in the workspace. After association with an Azure AI Foundry hub, Azure Container Registry and Application Insights resources cannot be disassociated (set to null).
115115

116116
You can use the Azure portal, Azure SDK/CLI options, or the infrastructure-as-code templates to update both Azure Application Insights and Azure Container Registry for the hub.
117117

0 commit comments

Comments
 (0)