You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-foundry/concepts/content-filtering.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -26,7 +26,7 @@ author: PatrickFarley
26
26
27
27
The content filtering system is powered by [Azure AI Content Safety](../../ai-services/content-safety/overview.md), and it works by running both the prompt input and completion output through a set of classification models designed to detect and prevent the output of harmful content. Variations in API configurations and application design might affect completions and thus filtering behavior.
28
28
29
-
With Azure OpenAI model deployments, you can use the default content filter or create your own content filter (described later on). Models available through **standard deployments** have content filtering enabled by default. To learn more about the default content filter enabled for standard deployments, see [Content safety for models curated by Azure AI in the model catalog](model-catalog-content-safety.md).
29
+
With Azure OpenAI model deployments, you can use the default content filter or create your own content filter (described later on). Models available through **standard deployments** have content filtering enabled by default. To learn more about the default content filter enabled for standard deployments, see [Content safety for Azure Direct Models](model-catalog-content-safety.md).
Copy file name to clipboardExpand all lines: articles/ai-foundry/concepts/foundry-models-overview.md
+34-38Lines changed: 34 additions & 38 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -37,14 +37,14 @@ Understanding the distinction between these categories helps you choose the righ
37
37
38
38
## Azure Direct Models
39
39
40
-
Azure Direct Models are models that are hosted and sold by Microsoft under Microsoft Product Terms. These models have undergone rigorous evaluation and are deeply integrated into Azure’s AI ecosystem. They offer enhanced integration, optimized performance, and direct Microsoft support, including enterprise-grade Service Level Agreements (SLAs).
40
+
Azure Direct Models are models that are hosted and sold by Microsoft under Microsoft Product Terms. These models have undergone rigorous evaluation and are deeply integrated into Azure's AI ecosystem. They offer enhanced integration, optimized performance, and direct Microsoft support, including enterprise-grade Service Level Agreements (SLAs).
41
41
42
42
Characteristics of Azure Direct Models:
43
43
44
44
- Official first-party support from Microsoft
45
45
- High level of integration with Azure services and infrastructure
46
46
- Extensive performance benchmarking and validation
47
-
- Adherence to Microsoft’s Responsible AI standards
47
+
- Adherence to Microsoft's Responsible AI standards
48
48
- Enterprise-grade scalability, reliability, and security
49
49
50
50
Azure Direct Models also have the benefit of flexible Provisioned Throughput, meaning you can use your quota and reservations across any of these models.
@@ -62,18 +62,18 @@ Characteristics of Azure Ecosystem Models:
62
62
63
63
Models are deployable as Managed Compute or Standard (pay-go) deployment options. The model provider selects how the models are deployable.
64
64
65
-
## Choosing Between Azure Direct and Azure Ecosystem Partner & Community Models
65
+
## Choosing between Azure Direct and Azure Ecosystem Models
66
66
67
67
68
68
When selecting models from Azure AI Foundry Models, consider the following:
69
-
* Use Case and Requirements: Azure Direct Models are ideal for scenarios requiring deep Azure integration, guaranteed support, and enterprise SLAs. Azure Ecosystem Models excel in specialized use cases and innovation-led scenarios.
70
-
* Support Expectations: Azure Direct Models come with robust Microsoft-provided support and maintenance. Azure Ecosystem Models are supported by their providers, with varying levels of SLA and support structures.
71
-
* Innovation and Specialization: Azure Ecosystem Models offer rapid access to specialized innovations and niche capabilities often developed by leading research labs and emerging AI providers.
69
+
***Use Case and Requirements**: Azure Direct Models are ideal for scenarios requiring deep Azure integration, guaranteed support, and enterprise SLAs. Azure Ecosystem Models excel in specialized use cases and innovation-led scenarios.
70
+
***Support Expectations**: Azure Direct Models come with robust Microsoft-provided support and maintenance. Azure Ecosystem Models are supported by their providers, with varying levels of SLA and support structures.
71
+
***Innovation and Specialization**: Azure Ecosystem Models offer rapid access to specialized innovations and niche capabilities often developed by leading research labs and emerging AI providers.
72
72
73
73
## Accessing Azure Ecosystem Models
74
74
75
75
Azure Ecosystem Models are accessible through Azure AI Foundry, providing:
76
-
* Comprehensive details about the model’s capabilities and integration requirements.
76
+
* Comprehensive details about the model's capabilities and integration requirements.
77
77
* Community ratings, usage data, and qualitative feedback to guide your decisions.
78
78
* Clear integration guidelines to help incorporate these models seamlessly into your Azure workflows.
79
79
@@ -85,48 +85,44 @@ Azure AI Foundry remains committed to providing a robust ecosystem, enabling cus
85
85
86
86
The model catalog organizes models into different collections:
87
87
88
-
***Curated by Azure AI**: The most popular partner models (open-weight and proprietary) packaged and optimized to work seamlessly on the Azure AI platform. Use of these models is subject to the model providers' license terms. When you deploy these models in Azure AI Foundry portal, their availability is subject to the applicable [Azure service-level agreement (SLA)](https://www.microsoft.com/licensing/docs/view/Service-Level-Agreements-SLA-for-Online-Services), and Microsoft provides support for deployment problems.
89
-
90
-
Models from partners such as Meta, NVIDIA, and Mistral AI are examples of models available in this collection on the catalog. You can identify these models by looking for a green checkmark on the model tiles in the catalog. Or you can filter by the **Curated by Azure AI** collection.
91
-
92
-
***Azure OpenAI models exclusively available on Azure**: Flagship Azure OpenAI models available through an integration with Azure OpenAI Service. Microsoft supports these models and their use according to the product terms and [SLA for Azure OpenAI Service](https://www.microsoft.com/licensing/docs/view/Service-Level-Agreements-SLA-for-Online-Services).
88
+
***Azure OpenAI models exclusively available on Azure**: Flagship Azure OpenAI models available through an integration with Azure OpenAI in Foundry Models. Microsoft supports these models and their use according to the product terms and [SLA for Azure OpenAI](https://www.microsoft.com/licensing/docs/view/Service-Level-Agreements-SLA-for-Online-Services).
93
89
94
90
***Open models from the Hugging Face hub**: Hundreds of models from the Hugging Face hub for real-time inference with managed compute. Hugging Face creates and maintains models listed in this collection. For help, use the [Hugging Face forum](https://discuss.huggingface.co) or [Hugging Face support](https://huggingface.co/support). Learn more in [Deploy open models with Azure AI Foundry](../how-to/deploy-models-managed.md).
95
91
96
92
You can submit a request to add a model to the model catalog by using [this form](https://forms.office.com/pages/responsepage.aspx?id=v4j5cvGGr0GRqy180BHbR_frVPkg_MhOoQxyrjmm7ZJUM09WNktBMURLSktOWEdDODBDRjg2NExKUy4u).
97
93
98
94
## Overview of Model Catalog capabilities
99
95
100
-
The model catalog in Azure AI Foundry portal is the hub to discover and use a wide range of models for building generative AI applications. The model catalog features hundreds of models across model providers such as Azure OpenAI Service, Mistral, Meta, Cohere, NVIDIA, and Hugging Face, including models that Microsoft trained. Models from providers other than Microsoft are Non-Microsoft Products as defined in [Microsoft Product Terms](https://www.microsoft.com/licensing/terms/welcome/welcomepage) and are subject to the terms provided with the models.
96
+
The model catalog in Azure AI Foundry portal is the hub to discover and use a wide range of models for building generative AI applications. The model catalog features hundreds of models across model providers such as Azure OpenAI, Mistral, Meta, Cohere, NVIDIA, and Hugging Face, including models that Microsoft trained. Models from providers other than Microsoft are Non-Microsoft Products as defined in [Microsoft Product Terms](https://www.microsoft.com/licensing/terms/welcome/welcomepage) and are subject to the terms provided with the models.
101
97
102
98
You can search and discover models that meet your need through keyword search and filters. Model catalog also offers the model performance leaderboard and benchmark metrics for select models. You can access them by selecting **Browse leaderboard** and **Compare Models**. Benchmark data is also accessible from the model card Benchmark tab.
103
99
104
-
On the model catalog filters, you’ll find:
105
-
106
-
* Collection: you can filter models based on the model provider collection.
107
-
* Industry: you can filter for the models that are trained on industry specific dataset.
108
-
* Capabilities: you can filter for unique model features such as reasoning and tool calling.
109
-
* Deployment options: you can filter for the models that support a specific deployment options.
110
-
* Standard: this option allows you to pay per API call.
111
-
* Provisioned: best suited for real-time scoring for large consistent volume.
112
-
* Batch: best suited for cost-optimized batch jobs, and not latency. No playground support is provided for the batch deployment.
113
-
* Managed compute: this option allows you to deploy a model on an Azure virtual machine. You will be billed for hosting and inferencing.
114
-
* Inference tasks: you can filter models based on the inference task type.
115
-
* Finetune tasks: you can filter models based on the finetune task type.
116
-
* Licenses: you can filter models based on the license type.
117
-
118
-
On the model card, you'll find:
119
-
120
-
* Quick facts: you will see key information about the model at a quick glance.
121
-
* Details: this page contains the detailed information about the model, including description, version info, supported data type, etc.
122
-
* Benchmarks: you will find performance benchmark metrics for select models.
123
-
* Existing deployments: if you have already deployed the model, you can find it under Existing deployments tab.
124
-
* License: you will find legal information related to model licensing.
125
-
* Artifacts: this tab will be displayed for open models only. You can see the model assets and download them via user interface.
100
+
On the **model catalog filters**, you'll find:
101
+
102
+
***Collection**: you can filter models based on the model provider collection.
103
+
***Industry**: you can filter for the models that are trained on industry specific dataset.
104
+
***Capabilities**: you can filter for unique model features such as reasoning and tool calling.
105
+
***Deployment options**: you can filter for the models that support a specific deployment options.
106
+
***Standard**: this option allows you to pay per API call.
107
+
***Provisioned**: best suited for real-time scoring for large consistent volume.
108
+
***Batch**: best suited for cost-optimized batch jobs, and not latency. No playground support is provided for the batch deployment.
109
+
***Managed compute**: this option allows you to deploy a model on an Azure virtual machine. You will be billed for hosting and inferencing.
110
+
***Inference tasks**: you can filter models based on the inference task type.
111
+
***Finetune tasks**: you can filter models based on the finetune task type.
112
+
***Licenses**: you can filter models based on the license type.
113
+
114
+
On the **model card**, you'll find:
115
+
116
+
***Quick facts**: you will see key information about the model at a quick glance.
117
+
***Details**: this page contains the detailed information about the model, including description, version info, supported data type, etc.
118
+
***Benchmarks**: you will find performance benchmark metrics for select models.
119
+
***Existing deployments**: if you have already deployed the model, you can find it under Existing deployments tab.
120
+
***License**: you will find legal information related to model licensing.
121
+
***Artifacts**: this tab will be displayed for open models only. You can see the model assets and download them via user interface.
126
122
127
123
## Model deployment: Managed compute and standard deployments
128
124
129
-
In addition to Azure OpenAI Service models, the model catalog offers two distinct ways to deploy models for your use: managed compute and standard deployments.
125
+
In addition to deploying to Azure OpenAI, the model catalog offers two distinct ways to deploy models for your use: managed compute and standard deployments.
130
126
131
127
The deployment options and features available for each model vary, as described in the following tables. [Learn more about data processing with the deployment options](../how-to/concept-data-privacy.md).
132
128
@@ -142,7 +138,7 @@ Network isolation | [Configure managed networks for Azure AI Foundry hubs](../ho
142
138
143
139
### Available models for supported deployment options
144
140
145
-
For Azure OpenAI models, see [Azure OpenAI Service Models](../../ai-services/openai/concepts/models.md).
141
+
For Azure OpenAI models, see [Azure OpenAI](../../ai-services/openai/concepts/models.md).
146
142
147
143
To view a list of supported models for standard deployment or Managed Compute, go to the home page of the model catalog in [Azure AI Foundry](https://ai.azure.com). Use the **Deployment options** filter to select either **Standard deployment** or **Managed Compute**.
148
144
@@ -156,7 +152,7 @@ To view a list of supported models for standard deployment or Managed Compute, g
156
152
## Model lifecycle: deprecation and retirement
157
153
AI models evolve fast, and when a new version or a new model with updated capabilities in the same model family become available, older models may be retired in the AI Foundry model catalog. To allow for a smooth transition to a newer model version, some models provide users with the option to enable automatic updates. To learn more about the model lifecycle of different models, upcoming model retirement dates, and suggested replacement models and versions, see:
158
154
159
-
-[Azure OpenAI Service model deprecations and retirements](../../ai-services/openai/concepts/model-retirements.md)
155
+
-[Azure OpenAI model deprecations and retirements](../../ai-services/openai/concepts/model-retirements.md)
160
156
-[Standard deployment model deprecations and retirements](../concepts/model-lifecycle-retirement.md)
Copy file name to clipboardExpand all lines: articles/ai-foundry/concepts/models-featured.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -250,8 +250,8 @@ See [the Microsoft model collection in Azure AI Foundry portal](https://ai.azure
250
250
251
251
Mistral AI offers two categories of models, namely:
252
252
253
-
-_Premium models_: These include Mistral Large, Mistral Small, Mistral-OCR-2503, and Ministral 3B models, and are available as standard deployments with serverless pay per token offer.
254
-
-_Open models_: These include Mistral-small-2503, Codestral, and Mistral Nemo (that are available as standard deployments with serverless pay per token offer), and [Mixtral-8x7B-Instruct-v01, Mixtral-8x7B-v01, Mistral-7B-Instruct-v01, and Mistral-7B-v01](../how-to/deploy-models-mistral-open.md)(that are available to download and run on self-hosted managed endpoints).
253
+
-_Premium models_: These include Mistral Large, Mistral Small, Mistral-OCR-2503, and Ministral 3B models, and are available as standard deployments.
254
+
-_Open models_: These include Mistral-small-2503, Codestral, and Mistral Nemo (that are available as standard deployments), and [Mixtral-8x7B-Instruct-v01, Mixtral-8x7B-v01, Mistral-7B-Instruct-v01, and Mistral-7B-v01](../how-to/deploy-models-mistral-open.md)(that are available to download and run on self-hosted managed endpoints).
255
255
256
256
257
257
| Model | Type | Capabilities |
@@ -364,5 +364,5 @@ For examples of how to use Stability AI models, see the following examples:
364
364
-[Deploy models as standard deployments](../how-to/deploy-models-serverless.md)
365
365
-[Model catalog and collections in Azure AI Foundry portal](../how-to/model-catalog-overview.md)
366
366
-[Region availability for models in standard deployments](../how-to/deploy-models-serverless-availability.md)
367
-
-[Content safety for models curated by Azure AI in the model catalog](model-catalog-content-safety.md)
367
+
-[Content safety for Azure Direct Models](model-catalog-content-safety.md)
Copy file name to clipboardExpand all lines: articles/ai-foundry/how-to/concept-data-privacy.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -35,7 +35,7 @@ Deploying models to managed compute deploys model weights to dedicated virtual m
35
35
36
36
You manage the infrastructure for these managed compute resources. Azure data, privacy, and security commitments apply. To learn more about Azure compliance offerings applicable to Azure AI Foundry, see the [Azure Compliance Offerings page](https://servicetrust.microsoft.com/DocumentPage/7adf2d9e-d7b5-4e71-bad8-713e6a183cf3).
37
37
38
-
Although containers for **Curated by Azure AI** models are scanned for vulnerabilities that could exfiltrate data, not all models available through the model catalog are scanned. To reduce the risk of data exfiltration, you can [help protect your deployment by using virtual networks](configure-managed-network.md). You can also use [Azure Policy](../../ai-services/policy-reference.md) to regulate the models that your users can deploy.
38
+
Although containers for **Azure Direct Models** are scanned for vulnerabilities that could exfiltrate data, not all models available through the model catalog are scanned. To reduce the risk of data exfiltration, you can [help protect your deployment by using virtual networks](configure-managed-network.md). You can also use [Azure Policy](../../ai-services/policy-reference.md) to regulate the models that your users can deploy.
39
39
40
40
:::image type="content" source="../media/explore/subscription-service-cycle.png" alt-text="Diagram that shows the platform service life cycle." lightbox="../media/explore/subscription-service-cycle.png":::
Copy file name to clipboardExpand all lines: articles/ai-foundry/how-to/configure-managed-network.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -29,7 +29,7 @@ You need to configure following network isolation configurations.
29
29
- Choose network isolation mode. You have two options: allow internet outbound mode or allow only approved outbound mode.
30
30
- If you use Visual Studio Code integration with allow only approved outbound mode, create FQDN outbound rules described in the [use Visual Studio Code](#scenario-use-visual-studio-code) section.
31
31
- If you use HuggingFace models in Models with allow only approved outbound mode, create FQDN outbound rules described in the [use HuggingFace models](#scenario-use-huggingface-models) section.
32
-
- If you use one of the open-source models with allow only approved outbound mode, create FQDN outbound rules described in the [curated by Azure AI](#scenario-curated-by-azure-ai) section.
32
+
- If you use one of the open-source models with allow only approved outbound mode, create FQDN outbound rules described in the [Azure Direct Models](#scenario-azure-direct-models) section.
33
33
34
34
## Network isolation architecture and isolation modes
35
35
@@ -759,7 +759,7 @@ For Azure AI Foundry to run with private networking, there are a set of required
759
759
|`BatchNodeManagement.region`| Outbound | Communication with Azure Batch back-end for Azure AI Foundry compute instances/clusters. |
760
760
|`AzureResourceManager`| Outbound | Creation of Azure resources with Azure AI Foundry, Azure CLI, and Azure AI Foundry SDK. |
761
761
|`AzureFrontDoor.FirstParty`| Outbound | Access docker images provided by Microsoft. |
762
-
|`MicrosoftContainerRegistry`| Outbound | Access docker images provided by Microsoft. Setup of the Azure AI Foundry router for Azure Kubernetes Service. |
762
+
|`MicrosoftContainerRegistry`| Outbound | Access docker images provided by Microsoft. Setup of the Azure AI Foundry router for Azure Kubernetes Service. |
763
763
|`AzureMonitor`| Outbound | Used to log monitoring and metrics to Azure Monitor. Only needed if you haven't secured Azure Monitor for the workspace. This outbound is also used to log information for support incidents. |
764
764
|`VirtualNetwork`| Outbound | Required when private endpoints are present in the virtual network or peered virtual networks. |
765
765
@@ -812,7 +812,7 @@ If you plan to use __HuggingFace models__ with the hub, add outbound _FQDN_ rule
812
812
* cnd.auth0.com
813
813
* cdn-lfs.huggingface.co
814
814
815
-
### Scenario: Curated by Azure AI
815
+
### Scenario: Azure Direct Models
816
816
817
817
These models involve dynamic installation of dependencies at runtime, and require outbound _FQDN_ rules to allow traffic to the following hosts:
0 commit comments