Skip to content

Commit c49e3a7

Browse files
authored
Merge pull request #4770 from sdgilley/sdg-merge-in-non-fdp
[DIRTY PR] merge non-FDP-features into release-build-ai-foundry
2 parents bad997f + b1d1c5c commit c49e3a7

File tree

299 files changed

+1128
-1016
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

299 files changed

+1128
-1016
lines changed

articles/ai-foundry/concepts/ai-red-teaming-agent.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,11 +17,11 @@ author: lgayhardt
1717

1818
The AI Red Teaming Agent (preview) is a powerful tool designed to help organizations proactively find safety risks associated with generative AI systems during design and development of generative AI models and applications.
1919

20-
Traditional red teaming involves exploiting the cyber kill chain and describes the process by which a system is tested for security vulnerabilities. However, with the rise of generative AI, the term AI red teaming has been coined to describe probing for novel risks (both content safety and security related) that these systems present and refers to simulating the behavior of an adversarial user who is trying to cause your AI system to misbehave in a particular way.
20+
Traditional red teaming involves exploiting the cyber kill chain and describes the process by which a system is tested for security vulnerabilities. However, with the rise of generative AI, the term AI red teaming has been coined to describe probing for novel risks (both content and security related) that these systems present and refers to simulating the behavior of an adversarial user who is trying to cause your AI system to misbehave in a particular way.
2121

2222
The AI Red Teaming Agent leverages Microsoft's open-source framework for Python Risk Identification Tool's ([PyRIT](https://github.com/Azure/PyRIT)) AI red teaming capabilities along with Azure AI Foundry's [Risk and Safety Evaluations](./evaluation-metrics-built-in.md#risk-and-safety-evaluators) to help you automatically assess safety issues in three ways:
2323

24-
- **Automated scans for content safety risks:** Firstly, you can automatically scan your model and application endpoints for safety risks by simulating adversarial probing.
24+
- **Automated scans for content risks:** Firstly, you can automatically scan your model and application endpoints for safety risks by simulating adversarial probing.
2525
- **Evaluate probing success:** Next, you can evaluate and score each attack-response pair to generate insightful metrics such as Attack Success Rate (ASR).
2626
- **Reporting and logging** Finally, you can generate a score card of the attack probing techniques and risk categories to help you decide if the system is ready for deployment. Findings can be logged, monitored, and tracked over time directly in Azure AI Foundry, ensuring compliance and continuous risk mitigation.
2727

articles/ai-foundry/concepts/ai-resources.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@ When a single platform team is responsible for the setup of cloud resources, the
3535

3636
## Set up and secure a hub for your team
3737

38-
Get started by [creating your first hub in Azure AI Foundry portal](../how-to/create-azure-ai-resource.md), or use [Azure portal](../how-to/create-secure-ai-hub.md) or [templates](../how-to/create-azure-ai-hub-template.md) for advanced configuration options. You can customize networking, identity, encryption, monitoring, or tags, to meet compliance with your organizations requirements.
38+
Get started by [creating your first hub in Azure AI Foundry portal](../how-to/create-azure-ai-resource.md), or use [Azure portal](../how-to/create-secure-ai-hub.md) or [templates](../how-to/create-azure-ai-hub-template.md) for advanced configuration options. You can customize networking, identity, encryption, monitoring, or tags, to meet compliance with your organization's requirements.
3939

4040
Often, projects in a business domain require access to the same company resources such as vector indices, model endpoints, or repos. As a team lead, you can preconfigure connectivity with these resources within a hub, so developers can access them from any new project workspace without delay on IT.
4141

@@ -79,7 +79,7 @@ Projects also have specific settings that only hold for that project:
7979
8080
## Azure AI services API access keys
8181

82-
The hub allows you to set up connections to existing Azure OpenAI or Azure AI Services resource types, which can be used to host model deployments. You can access these model deployments from connected resources in Azure AI Foundry portal. Keys to connected resources can be listed from the Azure AI Foundry portal or Azure portal. For more information, see [Find Azure AI Foundry resources in the Azure portal](#find-azure-ai-foundry-resources-in-the-azure-portal).
82+
The hub allows you to set up connections to existing Azure OpenAI or Azure AI Services resource types, which can be used to host model deployments. You can access these model deployments from connected resources in Azure AI Foundry portal. Keys to connected resources can be listed from the Azure AI Foundry portal or Azure portal. For more information, see [Find Azure AI Foundry Service in the Azure portal](#find-azure-ai-foundry-services-in-the-azure-portal).
8383

8484
### Virtual networking
8585

@@ -118,7 +118,7 @@ If you require to group costs of these different services together, we recommend
118118

119119
You can use [cost management](/azure/cost-management-billing/costs/quick-acm-cost-analysis) and [Azure resource tags](/azure/azure-resource-manager/management/tag-resources) to help with a detailed resource-level cost breakdown, or run [Azure pricing calculator](https://azure.microsoft.com/pricing/calculator/) on the above listed resources to obtain a pricing estimate. For more information, see [Plan and manage costs for Azure AI services](../how-to/costs-plan-manage.md).
120120

121-
## Find Azure AI Foundry resources in the Azure portal
121+
## Find Azure AI Foundry Services in the Azure portal
122122

123123
In the [Azure portal](https://portal.azure.com), search for and then select **Azure AI Foundry** entry. From the AI Foundry section of the portal, you can find your AI Foundry resources.
124124

articles/ai-foundry/concepts/architecture.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -31,12 +31,12 @@ At the top level, Azure AI Foundry provides access to the following resources:
3131

3232
For more information, visit [Azure OpenAI in Azure AI Foundry portal](../azure-openai-in-azure-ai-foundry.md).
3333

34-
- **Management center**: The management center streamlines governance and management of Azure AI Foundry resources such as hubs, projects, connected resources, and deployments.
34+
- **Management center**: The management center streamlines governance and management of Azure AI Foundry services such as hubs, projects, connected resources, and deployments.
3535

3636
For more information, visit [Management center](management-center.md).
3737
- **Azure AI Foundry hub**: The hub is the top-level resource in Azure AI Foundry portal, and is based on the Azure Machine Learning service. The Azure resource provider for a hub is `Microsoft.MachineLearningServices/workspaces`, and the kind of resource is `Hub`. It provides the following features:
3838
- Security configuration including a managed network that spans projects and model endpoints.
39-
- Compute resources for interactive development, fine-tuning, open source, and serverless model deployments.
39+
- Compute resources for interactive development, fine-tuning, open source, and standard deployment for models.
4040
- Connections to other Azure services such as Azure OpenAI, Azure AI services, and Azure AI Search. Hub-scoped connections are shared with projects created from the hub.
4141
- Project management. A hub can have multiple child projects.
4242
- An associated Azure storage account for data upload and artifact storage.
@@ -49,7 +49,7 @@ At the top level, Azure AI Foundry provides access to the following resources:
4949
- Project-scoped connections. For example, project members might need private access to data stored in an Azure Storage account without giving that same access to other projects.
5050
- Open source model deployments from catalog and fine-tuned model endpoints.
5151

52-
:::image type="content" source="../media/concepts/resource-provider-connected-resources.svg" alt-text="Diagram of the relationship between Azure AI Foundry resources." :::
52+
:::image type="content" source="../media/concepts/resource-provider-connected-resources.svg" alt-text="Diagram of the relationship between Azure AI Foundry services." :::
5353

5454
For more information, visit [Hubs and projects overview](ai-resources.md).
5555

articles/ai-foundry/concepts/content-filtering.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ author: PatrickFarley
2626

2727
The content filtering system is powered by [Azure AI Content Safety](../../ai-services/content-safety/overview.md), and it works by running both the prompt input and completion output through a set of classification models designed to detect and prevent the output of harmful content. Variations in API configurations and application design might affect completions and thus filtering behavior.
2828

29-
With Azure OpenAI model deployments, you can use the default content filter or create your own content filter (described later on). Models available through **serverless APIs** have content filtering enabled by default. To learn more about the default content filter enabled for serverless APIs, see [Content safety for models curated by Azure AI in the model catalog](model-catalog-content-safety.md).
29+
With Azure OpenAI model deployments, you can use the default content filter or create your own content filter (described later on). Models available through **serverless APIs** have content filtering enabled by default. To learn more about the default content filter enabled for serverless APIs, see [Guardrails & controls for Azure Direct Models in the model catalog](model-catalog-content-safety.md).
3030

3131
## Language support
3232

articles/ai-foundry/concepts/deployments-overview.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -25,9 +25,9 @@ Deployment options vary depending on the model offering:
2525

2626
Azure AI Foundry offers four different deployment options:
2727

28-
|Name | Azure OpenAI | Azure AI model inference | Standard deployment | Managed compute |
28+
|Name | Azure OpenAI | Azure AI Foundry Models | Standard deployment | Managed compute |
2929
|-------------------------------|----------------------|-------------------|----------------|-----------------|
30-
| Which models can be deployed? | [Azure OpenAI models](../../ai-services/openai/concepts/models.md) | [Azure OpenAI models and Standard deployment](../../ai-foundry/model-inference/concepts/models.md) | [Standard deployment](../how-to/model-catalog-overview.md#content-safety-for-models-deployed-via-serverless-apis) | [Open and custom models](../how-to/model-catalog-overview.md#availability-of-models-for-deployment-as-managed-compute) |
30+
| Which models can be deployed? | [Azure OpenAI models](../../ai-services/openai/concepts/models.md) | [Azure OpenAI models and Standard deployment](../../ai-foundry/model-inference/concepts/models.md) | [Standard deployment](../how-to/model-catalog-overview.md) | [Open and custom models](../how-to/model-catalog-overview.md#availability-of-models-for-deployment-as-managed-compute) |
3131
| Deployment resource | Azure OpenAI resource | Azure AI services resource | AI project resource | AI project resource |
3232
| Requires Hubs/Projects | No | No | Yes | Yes |
3333
| Data processing options | Regional <br /> Data-zone <br /> Global | Global | Regional | Regional |
@@ -37,7 +37,7 @@ Azure AI Foundry offers four different deployment options:
3737
| Key-less authentication | Yes | Yes | No | No |
3838
| Best suited when | You're planning to use only OpenAI models | You're planning to take advantage of the flagship models in Azure AI catalog, including OpenAI. | You're planning to use a single model from a specific provider (excluding OpenAI). | If you plan to use open models and you have enough compute quota available in your subscription. |
3939
| Billing bases | Token usage & [provisioned throughput units](../../ai-services/openai/concepts/provisioned-throughput.md) | Token usage | Token usage<sup>1</sup> | Compute core hours<sup>2</sup> |
40-
| Deployment instructions | [Deploy to Azure OpenAI](../how-to/deploy-models-openai.md) | [Deploy to Azure AI model inference](../model-inference/how-to/create-model-deployments.md) | [Deploy to Standard deployment](../how-to/deploy-models-serverless.md) | [Deploy to Managed compute](../how-to/deploy-models-managed.md) |
40+
| Deployment instructions | [Deploy to Azure OpenAI](../how-to/deploy-models-openai.md) | [Deploy to Foundry Models](../model-inference/how-to/create-model-deployments.md) | [Deploy to Standard deployment](../how-to/deploy-models-serverless.md) | [Deploy to Managed compute](../how-to/deploy-models-managed.md) |
4141

4242
<sup>1</sup> A minimal endpoint infrastructure is billed per minute. You aren't billed for the infrastructure that hosts the model in pay-as-you-go. After you delete the endpoint, no further charges accrue.
4343

@@ -50,7 +50,7 @@ Azure AI Foundry offers four different deployment options:
5050

5151
Azure AI Foundry encourages you to explore various deployment options and choose the one that best suites your business and technical needs. In general, Consider using the following approach to select a deployment option:
5252

53-
* Start with [Azure AI model inference](../../ai-foundry/model-inference/overview.md), which is the option with the largest scope. This option allows you to iterate and prototype faster in your application without having to rebuild your architecture each time you decide to change something. If you're using Azure AI Foundry hubs or projects, enable this option by [turning on the Azure AI model inference feature](../model-inference/how-to/quickstart-ai-project.md#configure-the-project-to-use-azure-ai-model-inference).
53+
* Start with [Foundry Models](../../ai-foundry/model-inference/overview.md), which is the option with the largest scope. This option allows you to iterate and prototype faster in your application without having to rebuild your architecture each time you decide to change something. If you're using Azure AI Foundry hubs or projects, enable this option by [turning on the Foundry Models feature](../model-inference/how-to/quickstart-ai-project.md#configure-the-project-to-use-foundry-models).
5454

5555
* When you're looking to use a specific model:
5656

@@ -63,8 +63,8 @@ Azure AI Foundry encourages you to explore various deployment options and choose
6363

6464
## Related content
6565

66-
* [Configure your AI project to use Azure AI model inference](../../ai-foundry/model-inference/how-to/quickstart-ai-project.md)
67-
* [Add and configure models to Azure AI model inference](../model-inference/how-to/create-model-deployments.md)
66+
* [Configure your AI project to use Foundry Models](../../ai-foundry/model-inference/how-to/quickstart-ai-project.md)
67+
* [Add and configure models to Foundry Models](../model-inference/how-to/create-model-deployments.md)
6868
* [Deploy Azure OpenAI models with Azure AI Foundry](../how-to/deploy-models-openai.md)
6969
* [Deploy open models with Azure AI Foundry](../how-to/deploy-models-managed.md)
7070
* [Model catalog and collections in Azure AI Foundry portal](../how-to/model-catalog-overview.md)

0 commit comments

Comments
 (0)