Skip to content

Commit c4397bd

Browse files
authored
Merge branch 'release-build-ai-foundry' into santiagxf/foundry-models
2 parents c9aa1e0 + d6e725d commit c4397bd

File tree

86 files changed

+162
-162
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

86 files changed

+162
-162
lines changed

articles/ai-foundry/concepts/content-filtering.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ author: PatrickFarley
2626

2727
The content filtering system is powered by [Azure AI Content Safety](../../ai-services/content-safety/overview.md), and it works by running both the prompt input and completion output through a set of classification models designed to detect and prevent the output of harmful content. Variations in API configurations and application design might affect completions and thus filtering behavior.
2828

29-
With Azure OpenAI model deployments, you can use the default content filter or create your own content filter (described later on). Models available through **standard deployments** have content filtering enabled by default. To learn more about the default content filter enabled for standard deployments, see [Content safety for models curated by Azure AI in the model catalog](model-catalog-content-safety.md).
29+
With Azure OpenAI model deployments, you can use the default content filter or create your own content filter (described later on). Models available through **standard deployments** have content filtering enabled by default. To learn more about the default content filter enabled for standard deployments, see [Content safety for Azure Direct Models](model-catalog-content-safety.md).
3030

3131
## Language support
3232

articles/ai-foundry/concepts/foundry-models-overview.md

Lines changed: 34 additions & 38 deletions
Original file line numberDiff line numberDiff line change
@@ -37,14 +37,14 @@ Understanding the distinction between these categories helps you choose the righ
3737

3838
## Azure Direct Models
3939

40-
Azure Direct Models are models that are hosted and sold by Microsoft under Microsoft Product Terms. These models have undergone rigorous evaluation and are deeply integrated into Azures AI ecosystem. They offer enhanced integration, optimized performance, and direct Microsoft support, including enterprise-grade Service Level Agreements (SLAs).
40+
Azure Direct Models are models that are hosted and sold by Microsoft under Microsoft Product Terms. These models have undergone rigorous evaluation and are deeply integrated into Azure's AI ecosystem. They offer enhanced integration, optimized performance, and direct Microsoft support, including enterprise-grade Service Level Agreements (SLAs).
4141

4242
Characteristics of Azure Direct Models:
4343

4444
- Official first-party support from Microsoft
4545
- High level of integration with Azure services and infrastructure
4646
- Extensive performance benchmarking and validation
47-
- Adherence to Microsofts Responsible AI standards
47+
- Adherence to Microsoft's Responsible AI standards
4848
- Enterprise-grade scalability, reliability, and security
4949

5050
Azure Direct Models also have the benefit of flexible Provisioned Throughput, meaning you can use your quota and reservations across any of these models.
@@ -62,18 +62,18 @@ Characteristics of Azure Ecosystem Models:
6262

6363
Models are deployable as Managed Compute or Standard (pay-go) deployment options. The model provider selects how the models are deployable.
6464

65-
## Choosing Between Azure Direct and Azure Ecosystem Partner & Community Models
65+
## Choosing between Azure Direct and Azure Ecosystem Models
6666

6767

6868
When selecting models from Azure AI Foundry Models, consider the following:
69-
* Use Case and Requirements: Azure Direct Models are ideal for scenarios requiring deep Azure integration, guaranteed support, and enterprise SLAs. Azure Ecosystem Models excel in specialized use cases and innovation-led scenarios.
70-
* Support Expectations: Azure Direct Models come with robust Microsoft-provided support and maintenance. Azure Ecosystem Models are supported by their providers, with varying levels of SLA and support structures.
71-
* Innovation and Specialization: Azure Ecosystem Models offer rapid access to specialized innovations and niche capabilities often developed by leading research labs and emerging AI providers.
69+
* **Use Case and Requirements**: Azure Direct Models are ideal for scenarios requiring deep Azure integration, guaranteed support, and enterprise SLAs. Azure Ecosystem Models excel in specialized use cases and innovation-led scenarios.
70+
* **Support Expectations**: Azure Direct Models come with robust Microsoft-provided support and maintenance. Azure Ecosystem Models are supported by their providers, with varying levels of SLA and support structures.
71+
* **Innovation and Specialization**: Azure Ecosystem Models offer rapid access to specialized innovations and niche capabilities often developed by leading research labs and emerging AI providers.
7272

7373
## Accessing Azure Ecosystem Models
7474

7575
Azure Ecosystem Models are accessible through Azure AI Foundry, providing:
76-
* Comprehensive details about the models capabilities and integration requirements.
76+
* Comprehensive details about the model's capabilities and integration requirements.
7777
* Community ratings, usage data, and qualitative feedback to guide your decisions.
7878
* Clear integration guidelines to help incorporate these models seamlessly into your Azure workflows.
7979

@@ -85,48 +85,44 @@ Azure AI Foundry remains committed to providing a robust ecosystem, enabling cus
8585

8686
The model catalog organizes models into different collections:
8787

88-
* **Curated by Azure AI**: The most popular partner models (open-weight and proprietary) packaged and optimized to work seamlessly on the Azure AI platform. Use of these models is subject to the model providers' license terms. When you deploy these models in Azure AI Foundry portal, their availability is subject to the applicable [Azure service-level agreement (SLA)](https://www.microsoft.com/licensing/docs/view/Service-Level-Agreements-SLA-for-Online-Services), and Microsoft provides support for deployment problems.
89-
90-
Models from partners such as Meta, NVIDIA, and Mistral AI are examples of models available in this collection on the catalog. You can identify these models by looking for a green checkmark on the model tiles in the catalog. Or you can filter by the **Curated by Azure AI** collection.
91-
92-
* **Azure OpenAI models exclusively available on Azure**: Flagship Azure OpenAI models available through an integration with Azure OpenAI Service. Microsoft supports these models and their use according to the product terms and [SLA for Azure OpenAI Service](https://www.microsoft.com/licensing/docs/view/Service-Level-Agreements-SLA-for-Online-Services).
88+
* **Azure OpenAI models exclusively available on Azure**: Flagship Azure OpenAI models available through an integration with Azure OpenAI in Foundry Models. Microsoft supports these models and their use according to the product terms and [SLA for Azure OpenAI](https://www.microsoft.com/licensing/docs/view/Service-Level-Agreements-SLA-for-Online-Services).
9389

9490
* **Open models from the Hugging Face hub**: Hundreds of models from the Hugging Face hub for real-time inference with managed compute. Hugging Face creates and maintains models listed in this collection. For help, use the [Hugging Face forum](https://discuss.huggingface.co) or [Hugging Face support](https://huggingface.co/support). Learn more in [Deploy open models with Azure AI Foundry](../how-to/deploy-models-managed.md).
9591

9692
You can submit a request to add a model to the model catalog by using [this form](https://forms.office.com/pages/responsepage.aspx?id=v4j5cvGGr0GRqy180BHbR_frVPkg_MhOoQxyrjmm7ZJUM09WNktBMURLSktOWEdDODBDRjg2NExKUy4u).
9793

9894
## Overview of Model Catalog capabilities
9995

100-
The model catalog in Azure AI Foundry portal is the hub to discover and use a wide range of models for building generative AI applications. The model catalog features hundreds of models across model providers such as Azure OpenAI Service, Mistral, Meta, Cohere, NVIDIA, and Hugging Face, including models that Microsoft trained. Models from providers other than Microsoft are Non-Microsoft Products as defined in [Microsoft Product Terms](https://www.microsoft.com/licensing/terms/welcome/welcomepage) and are subject to the terms provided with the models.
96+
The model catalog in Azure AI Foundry portal is the hub to discover and use a wide range of models for building generative AI applications. The model catalog features hundreds of models across model providers such as Azure OpenAI, Mistral, Meta, Cohere, NVIDIA, and Hugging Face, including models that Microsoft trained. Models from providers other than Microsoft are Non-Microsoft Products as defined in [Microsoft Product Terms](https://www.microsoft.com/licensing/terms/welcome/welcomepage) and are subject to the terms provided with the models.
10197

10298
You can search and discover models that meet your need through keyword search and filters. Model catalog also offers the model performance leaderboard and benchmark metrics for select models. You can access them by selecting **Browse leaderboard** and **Compare Models**. Benchmark data is also accessible from the model card Benchmark tab.
10399

104-
On the model catalog filters, youll find:
105-
106-
* Collection: you can filter models based on the model provider collection.
107-
* Industry: you can filter for the models that are trained on industry specific dataset.
108-
* Capabilities: you can filter for unique model features such as reasoning and tool calling.
109-
* Deployment options: you can filter for the models that support a specific deployment options.
110-
* Standard: this option allows you to pay per API call.
111-
* Provisioned: best suited for real-time scoring for large consistent volume.
112-
* Batch: best suited for cost-optimized batch jobs, and not latency. No playground support is provided for the batch deployment.
113-
* Managed compute: this option allows you to deploy a model on an Azure virtual machine. You will be billed for hosting and inferencing.
114-
* Inference tasks: you can filter models based on the inference task type.
115-
* Finetune tasks: you can filter models based on the finetune task type.
116-
* Licenses: you can filter models based on the license type.
117-
118-
On the model card, you'll find:
119-
120-
* Quick facts: you will see key information about the model at a quick glance.
121-
* Details: this page contains the detailed information about the model, including description, version info, supported data type, etc.
122-
* Benchmarks: you will find performance benchmark metrics for select models.
123-
* Existing deployments: if you have already deployed the model, you can find it under Existing deployments tab.
124-
* License: you will find legal information related to model licensing.
125-
* Artifacts: this tab will be displayed for open models only. You can see the model assets and download them via user interface.
100+
On the **model catalog filters**, you'll find:
101+
102+
* **Collection**: you can filter models based on the model provider collection.
103+
* **Industry**: you can filter for the models that are trained on industry specific dataset.
104+
* **Capabilities**: you can filter for unique model features such as reasoning and tool calling.
105+
* **Deployment options**: you can filter for the models that support a specific deployment options.
106+
* **Standard**: this option allows you to pay per API call.
107+
* **Provisioned**: best suited for real-time scoring for large consistent volume.
108+
* **Batch**: best suited for cost-optimized batch jobs, and not latency. No playground support is provided for the batch deployment.
109+
* **Managed compute**: this option allows you to deploy a model on an Azure virtual machine. You will be billed for hosting and inferencing.
110+
* **Inference tasks**: you can filter models based on the inference task type.
111+
* **Finetune tasks**: you can filter models based on the finetune task type.
112+
* **Licenses**: you can filter models based on the license type.
113+
114+
On the **model card**, you'll find:
115+
116+
* **Quick facts**: you will see key information about the model at a quick glance.
117+
* **Details**: this page contains the detailed information about the model, including description, version info, supported data type, etc.
118+
* **Benchmarks**: you will find performance benchmark metrics for select models.
119+
* **Existing deployments**: if you have already deployed the model, you can find it under Existing deployments tab.
120+
* **License**: you will find legal information related to model licensing.
121+
* **Artifacts**: this tab will be displayed for open models only. You can see the model assets and download them via user interface.
126122

127123
## Model deployment: Managed compute and standard deployments
128124

129-
In addition to Azure OpenAI Service models, the model catalog offers two distinct ways to deploy models for your use: managed compute and standard deployments.
125+
In addition to deploying to Azure OpenAI, the model catalog offers two distinct ways to deploy models for your use: managed compute and standard deployments.
130126

131127
The deployment options and features available for each model vary, as described in the following tables. [Learn more about data processing with the deployment options](../how-to/concept-data-privacy.md).
132128

@@ -142,7 +138,7 @@ Network isolation | [Configure managed networks for Azure AI Foundry hubs](../ho
142138

143139
### Available models for supported deployment options
144140

145-
For Azure OpenAI models, see [Azure OpenAI Service Models](../../ai-services/openai/concepts/models.md).
141+
For Azure OpenAI models, see [Azure OpenAI](../../ai-services/openai/concepts/models.md).
146142

147143
To view a list of supported models for standard deployment or Managed Compute, go to the home page of the model catalog in [Azure AI Foundry](https://ai.azure.com). Use the **Deployment options** filter to select either **Standard deployment** or **Managed Compute**.
148144

@@ -156,7 +152,7 @@ To view a list of supported models for standard deployment or Managed Compute, g
156152
## Model lifecycle: deprecation and retirement
157153
AI models evolve fast, and when a new version or a new model with updated capabilities in the same model family become available, older models may be retired in the AI Foundry model catalog. To allow for a smooth transition to a newer model version, some models provide users with the option to enable automatic updates. To learn more about the model lifecycle of different models, upcoming model retirement dates, and suggested replacement models and versions, see:
158154

159-
- [Azure OpenAI Service model deprecations and retirements](../../ai-services/openai/concepts/model-retirements.md)
155+
- [Azure OpenAI model deprecations and retirements](../../ai-services/openai/concepts/model-retirements.md)
160156
- [Standard deployment model deprecations and retirements](../concepts/model-lifecycle-retirement.md)
161157

162158
## Managed compute

articles/ai-foundry/concepts/models-featured.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -250,8 +250,8 @@ See [the Microsoft model collection in Azure AI Foundry portal](https://ai.azure
250250

251251
Mistral AI offers two categories of models, namely:
252252

253-
- _Premium models_: These include Mistral Large, Mistral Small, Mistral-OCR-2503, and Ministral 3B models, and are available as standard deployments with serverless pay per token offer.
254-
- _Open models_: These include Mistral-small-2503, Codestral, and Mistral Nemo (that are available as standard deployments with serverless pay per token offer), and [Mixtral-8x7B-Instruct-v01, Mixtral-8x7B-v01, Mistral-7B-Instruct-v01, and Mistral-7B-v01](../how-to/deploy-models-mistral-open.md)(that are available to download and run on self-hosted managed endpoints).
253+
- _Premium models_: These include Mistral Large, Mistral Small, Mistral-OCR-2503, and Ministral 3B models, and are available as standard deployments.
254+
- _Open models_: These include Mistral-small-2503, Codestral, and Mistral Nemo (that are available as standard deployments), and [Mixtral-8x7B-Instruct-v01, Mixtral-8x7B-v01, Mistral-7B-Instruct-v01, and Mistral-7B-v01](../how-to/deploy-models-mistral-open.md)(that are available to download and run on self-hosted managed endpoints).
255255

256256

257257
| Model | Type | Capabilities |
@@ -364,5 +364,5 @@ For examples of how to use Stability AI models, see the following examples:
364364
- [Deploy models as standard deployments](../how-to/deploy-models-serverless.md)
365365
- [Model catalog and collections in Azure AI Foundry portal](../how-to/model-catalog-overview.md)
366366
- [Region availability for models in standard deployments](../how-to/deploy-models-serverless-availability.md)
367-
- [Content safety for models curated by Azure AI in the model catalog](model-catalog-content-safety.md)
367+
- [Content safety for Azure Direct Models](model-catalog-content-safety.md)
368368

articles/ai-foundry/how-to/concept-data-privacy.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@ Deploying models to managed compute deploys model weights to dedicated virtual m
3535

3636
You manage the infrastructure for these managed compute resources. Azure data, privacy, and security commitments apply. To learn more about Azure compliance offerings applicable to Azure AI Foundry, see the [Azure Compliance Offerings page](https://servicetrust.microsoft.com/DocumentPage/7adf2d9e-d7b5-4e71-bad8-713e6a183cf3).
3737

38-
Although containers for **Curated by Azure AI** models are scanned for vulnerabilities that could exfiltrate data, not all models available through the model catalog are scanned. To reduce the risk of data exfiltration, you can [help protect your deployment by using virtual networks](configure-managed-network.md). You can also use [Azure Policy](../../ai-services/policy-reference.md) to regulate the models that your users can deploy.
38+
Although containers for **Azure Direct Models** are scanned for vulnerabilities that could exfiltrate data, not all models available through the model catalog are scanned. To reduce the risk of data exfiltration, you can [help protect your deployment by using virtual networks](configure-managed-network.md). You can also use [Azure Policy](../../ai-services/policy-reference.md) to regulate the models that your users can deploy.
3939

4040
:::image type="content" source="../media/explore/subscription-service-cycle.png" alt-text="Diagram that shows the platform service life cycle." lightbox="../media/explore/subscription-service-cycle.png":::
4141

articles/ai-foundry/how-to/configure-managed-network.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@ You need to configure following network isolation configurations.
2929
- Choose network isolation mode. You have two options: allow internet outbound mode or allow only approved outbound mode.
3030
- If you use Visual Studio Code integration with allow only approved outbound mode, create FQDN outbound rules described in the [use Visual Studio Code](#scenario-use-visual-studio-code) section.
3131
- If you use HuggingFace models in Models with allow only approved outbound mode, create FQDN outbound rules described in the [use HuggingFace models](#scenario-use-huggingface-models) section.
32-
- If you use one of the open-source models with allow only approved outbound mode, create FQDN outbound rules described in the [curated by Azure AI](#scenario-curated-by-azure-ai) section.
32+
- If you use one of the open-source models with allow only approved outbound mode, create FQDN outbound rules described in the [Azure Direct Models](#scenario-azure-direct-models) section.
3333

3434
## Network isolation architecture and isolation modes
3535

@@ -759,7 +759,7 @@ For Azure AI Foundry to run with private networking, there are a set of required
759759
| `BatchNodeManagement.region` | Outbound | Communication with Azure Batch back-end for Azure AI Foundry compute instances/clusters. |
760760
| `AzureResourceManager` | Outbound | Creation of Azure resources with Azure AI Foundry, Azure CLI, and Azure AI Foundry SDK. |
761761
| `AzureFrontDoor.FirstParty` | Outbound | Access docker images provided by Microsoft. |
762-
| `MicrosoftContainerRegistry` | Outbound | Access docker images provided by Microsoft. Setup of the Azure AI Foundry router for Azure Kubernetes Service. |
762+
| `MicrosoftContainerRegistry` | Outbound | Access docker images provided by Microsoft. Setup of the Azure AI Foundry router for Azure Kubernetes Service. |
763763
| `AzureMonitor` | Outbound | Used to log monitoring and metrics to Azure Monitor. Only needed if you haven't secured Azure Monitor for the workspace. This outbound is also used to log information for support incidents. |
764764
| `VirtualNetwork` | Outbound | Required when private endpoints are present in the virtual network or peered virtual networks. |
765765

@@ -812,7 +812,7 @@ If you plan to use __HuggingFace models__ with the hub, add outbound _FQDN_ rule
812812
* cnd.auth0.com
813813
* cdn-lfs.huggingface.co
814814

815-
### Scenario: Curated by Azure AI
815+
### Scenario: Azure Direct Models
816816

817817
These models involve dynamic installation of dependencies at runtime, and require outbound _FQDN_ rules to allow traffic to the following hosts:
818818

0 commit comments

Comments
 (0)