You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-foundry/ai-services/content-safety-overview.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -46,4 +46,4 @@ Refer to the [Content Safety overview](/azure/ai-services/content-safety/overvie
46
46
47
47
## Next step
48
48
49
-
Get started using Azure AI Content Safety in [Azure AI Foundry portal](https://ai.azure.com) by following the [How-to guide](./how-to/content-safety.md).
49
+
Get started using Azure AI Content Safety in [Azure AI Foundry portal](https://ai.azure.com) by following the [How-to guide](/azure/ai-services/content-safety/how-to/foundry).
Copy file name to clipboardExpand all lines: articles/ai-foundry/ai-services/how-to/connect-azure-openai.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -100,7 +100,7 @@ Here are a few guides to help you get started with Azure OpenAI Service playgrou
100
100
-[Quickstart: Use the chat playground](../../quickstarts/get-started-playground.md)
101
101
-[Quickstart: Get started using Azure OpenAI Assistants](../../../ai-services/openai/assistants-quickstart.md?context=/azure/ai-studio/context/context)
102
102
-[Quickstart: Use GPT-4o in the real-time audio playground](../../../ai-services/openai/realtime-audio-quickstart.md?context=/azure/ai-studio/context/context)
103
-
-[Quickstart: Analyze images and video in the chat playground](../../quickstarts/multimodal-vision.md)
103
+
-[Quickstart: Analyze images and video in the chat playground](/azure/ai-services/openai/gpt-v-quickstart)
104
104
105
105
Each playground has different model requirements and capabilities. The supported regions vary depending on the model. For more information about model availability per region, see the [Azure OpenAI Service models documentation](../../../ai-services/openai/concepts/models.md).
106
106
@@ -145,5 +145,5 @@ At some point, you want to develop apps with code. Here are some developer resou
145
145
146
146
## Related content
147
147
148
-
-[Azure OpenAI in Azure AI Foundry portal](../../azure-openai-in-ai-studio.md)
148
+
-[Azure OpenAI in Azure AI Foundry portal](../../azure-openai-in-ai-foundry.md)
149
149
-[Use Azure AI services resources](./connect-ai-services.md)
Copy file name to clipboardExpand all lines: articles/ai-foundry/concepts/a-b-experimentation.md
+7-11Lines changed: 7 additions & 11 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,14 +1,15 @@
1
1
---
2
2
title: A/B experiments for AI applications
3
+
titleSuffix: Azure AI Foundry
3
4
description: Learn about conducting A/B experiments for AI applications.
4
-
author: s-polly
5
-
ms.author: scottpolly
5
+
author: lgayhardt
6
+
ms.author: lagayhar
6
7
ms.reviewer: skohlmeier
7
8
ms.service: azure-ai-foundry
8
9
ms.custom:
9
10
- ignite-2024
10
11
ms.topic: concept-article
11
-
ms.date: 11/22/2024
12
+
ms.date: 02/27/2025
12
13
13
14
#CustomerIntent: As an AI application developer, I want to learn about A/B experiments so that I can evaluate and improve my applications.
14
15
---
@@ -18,23 +19,21 @@ ms.date: 11/22/2024
18
19
> [!IMPORTANT]
19
20
>Items marked (preview) in this article are currently in public or private preview. This preview is provided without a service-level agreement, and we don't recommend it for production workloads. Certain features might not be supported or might have constrained capabilities. For more information, see [Supplemental Terms of Use for Microsoft Azure Previews](https://azure.microsoft.com/support/legal/preview-supplemental-terms/).
20
21
21
-
In the field of AI application development, A/B experimentation has emerged as a critical practice. It allows for continuous evaluation of AI applications, balancing business impact, risk, and cost. While offline and online evaluations provide some insights, they need to be supplemented with A/B experimentation to ensure the use of right metrics for measuring success. A/B experimentation involves comparing two versions of a feature, prompt, or model using feature flags or dynamic configuration to determine which performs better. This method is essential for several reasons:
22
+
In the field of AI application development, A/B experimentation has emerged as a critical practice. It allows for continuous evaluation of AI applications, balancing business impact, risk, and cost. While offline and online evaluations provide some insights, they need to be supplemented with A/B experimentation to ensure the use of right metrics for measuring success. A/B experimentation involves comparing two versions of a feature, prompt, or model using feature flags or dynamic configuration to determine which performs better. This method is essential for several reasons:
22
23
23
24
-**Enhancing Model Performance** - A/B experimentation allows developers to systematically test different versions of AI models, algorithms, or features to identify the most effective version. With controlled experiments, you can measure the effect of changes on key performance metrics, such as accuracy, user engagement, and response time. This iterative process enables you to identify the best model, helps fine-tuning and ensures that your models deliver the best possible results.
24
25
-**Reducing Bias and Improving Fairness** - AI models can inadvertently introduce biases, leading to unfair outcomes. A/B experimentation helps identify and mitigate these biases by comparing the performance of different model versions across diverse user groups. This ensures that the AI applications are fair and equitable, providing consistent performance for all users.
25
26
-**Accelerating Innovation** - A/B experimentation fosters a culture of innovation by encouraging continuous experimentation and learning. You can quickly validate new ideas and features, reducing the time and resources spent on unproductive approaches. This accelerates the development cycle and allows teams to bring innovative AI solutions to market faster.
26
27
-**Optimizing User Experience** - User experience is paramount in AI applications. A/B experimentation enables you to experiment with different user interface designs, interaction patterns, and personalization strategies. By analyzing user feedback and behavior, you can optimize the user experience, making AI applications more intuitive and engaging.
27
28
-**Data-Driven Decision Making** - A/B experimentation provides a robust framework for data-driven decision making. Instead of relying on intuition or assumptions, you can base your decisions on empirical evidence. This leads to more informed and effective strategies for improving AI applications.
28
29
29
-
30
30
## How does A/B experimentation fit into the AI application lifecycle?
31
31
32
-
33
32
A/B experimentation and offline evaluation are both essential components in the development of AI applications, each serving unique purposes that complement each other.
34
33
35
34
Offline evaluation involves testing AI models using test datasets to measure their performance on various metrics such as fluency and coherence. After selecting a model in the Azure AI Model Catalog or GitHub Model marketplace, offline preproduction evaluation is crucial for initial model validation during integration testing, allowing you to identify potential issues and make improvements before deploying the model or application to production.
36
35
37
-
However, offline evaluation has its limitations. It can't fully capture the complex interactions that occur in real-world scenarios. This is where A/B experimentation comes into play. By deploying different versions of the AI model or UX features to live users, A/B experimentation provides insights into how the model and application performs in real-world conditions. This helps you understand user behavior, identify unforeseen issues, and measure the impact of changes on model evaluation metrics, operational metrics (for example, latency) and business metrics (for example, account sign-ups, conversions, etc.).
36
+
However, offline evaluation has its limitations. It can't fully capture the complex interactions that occur in real-world scenarios. This is where A/B experimentation comes into play. By deploying different versions of the AI model or UX features to live users, A/B experimentation provides insights into how the model and application performs in real-world conditions. This helps you understand user behavior, identify unforeseen issues, and measure the impact of changes on model evaluation metrics, operational metrics (for example, latency), and business metrics (for example, account sign-ups, conversions, etc.).
38
37
39
38
As shown in the diagram, while offline evaluation is essential for initial model validation and refinement, A/B experimentation provides the real-world testing needed to ensure the AI application performs effectively and fairly in practice. Together, they form a comprehensive approach to developing robust, safe, and user-friendly AI applications.
40
39
@@ -52,8 +51,7 @@ We're significantly simplifying the evaluation and A/B experimentation process w
52
51
53
52
## Azure AI Partners
54
53
55
-
56
-
You're also welcome to use your own A/B experimentation provider to run experiments on your AI applications. There are several solutions to choose from available in the Azure Marketplace:
54
+
You're also welcome to use your own A/B experimentation provider to run experiments on your AI applications. There are several solutions to choose from available in Azure Marketplace:
57
55
58
56
### Statsig
59
57
@@ -67,8 +65,6 @@ You're also welcome to use your own A/B experimentation provider to run experime
67
65
### LaunchDarkly
68
66
[LaunchDarkly](https://azuremarketplace.microsoft.com/marketplace/apps/aad.launchdarkly?tab=Overview) is a feature management and experimentation platform built with software developers in mind. It enables you to manage feature flags on a large scale, run A/B tests and experiments, and progressively deliver software to ship with confidence.
Copy file name to clipboardExpand all lines: articles/ai-foundry/concepts/architecture.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -28,7 +28,7 @@ At the top level, Azure AI Foundry provides access to the following resources:
28
28
29
29
When you use Azure AI Foundry portal, you can directly work with Azure OpenAI without an Azure Studio project. Or you can use Azure OpenAI through a project.
30
30
31
-
For more information, visit [Azure OpenAI in Azure AI Foundry portal](../azure-openai-in-ai-studio.md).
31
+
For more information, visit [Azure OpenAI in Azure AI Foundry portal](../azure-openai-in-ai-foundry.md).
32
32
33
33
-**Management center**: The management center streamlines governance and management of Azure AI Foundry resources such as hubs, projects, connected resources, and deployments.
34
34
@@ -108,7 +108,7 @@ Azure AI services including Azure OpenAI provide control plane endpoints for ope
108
108
109
109
To reduce the complexity of Azure RBAC management, Azure AI Foundry provides a *control plane proxy* that allows you to perform operations on connected Azure AI services and Azure OpenAI resources. Performing operations on these resources through the control plane proxy only requires Azure RBAC permissions on the hub. The Azure AI Foundry service then performs the call to the Azure AI services or Azure OpenAI control plane endpoint on your behalf.
110
110
111
-
For more information, see [Role-based access control in Azure AI Foundry portal](rbac-ai-studio.md).
111
+
For more information, see [Role-based access control in Azure AI Foundry portal](rbac-ai-foundry.md).
0 commit comments