You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
While Azure AI Evaluation SDK supports running evaluations locally on your own machine, you might want to delegate the job remotely to the cloud. For example, after you ran local evaluations on small test data to help assess your generative AI application prototypes, now you move into pre-deployment testing and need run evaluations on a large dataset. Cloud evaluation frees you from managing your local compute infrastructure, and enables you to integrate evaluations as tests into your CI/CD pipelines. After deployment, you might want to [continuously evaluate](../online-evaluation.md) your applications for post-deployment monitoring.
20
+
The Azure AI Evaluation SDK supports running evaluations locally on your own machine and also in the cloud. For example, after you ran local evaluations on small test data to help assess your generative AI application prototypes, you can move into pre-deployment testing and run evaluations on a large dataset. Evaluating your applications in the cloud frees you from managing your local compute infrastructure, and enables you to integrate evaluations as tests into your continuous integration and continuous delivery (CI/CD) pipelines. After deployment, you can choose to [continuously evaluate](../online-evaluation.md) your applications for post-deployment monitoring.
21
21
22
-
In this article, you learn how to run evaluations in the cloud (preview) in pre-deployment testing on a test dataset. Using the Azure AI Projects SDK, you'll have evaluation results automatically logged into your Azure AI project for better observability. This feature supports all Microsoft curated [built-in evaluators](../../concepts/observability.md#what-are-evaluators) and your own [custom evaluators](../../concepts/evaluation-evaluators/custom-evaluators.md) which can be located in the [Evaluator library](../evaluate-generative-ai-app.md#view-and-manage-the-evaluators-in-the-evaluator-library) and have the same project-scope RBAC.
22
+
In this article, you learn how to run evaluations in the cloud (preview) in pre-deployment testing on a test dataset. When you use the Azure AI Projects SDK, evaluation results are automatically logged into your Azure AI project for better observability. This feature supports all Microsoft curated [built-in evaluators](../../concepts/observability.md#what-are-evaluators) and your own [custom evaluators](../../concepts/evaluation-evaluators/custom-evaluators.md). Your evaluators can be located in the [Evaluator library](../evaluate-generative-ai-app.md#view-and-manage-the-evaluators-in-the-evaluator-library) and have the same project-scope role-based access control (RBAC).
23
23
24
-
## Prerequisite set up steps for Azure AI Foundry Projects
24
+
## Prerequisites
25
25
26
-
- Azure AI Foundry project in the same supported [regions](../../concepts/evaluation-evaluators/risk-safety-evaluators.md#azure-ai-foundry-project-configuration-and-region-support) as risk and safety evaluators (preview). If you don't have an existing project, follow the guide [How to create Azure AI Foundry project](../create-projects.md?tabs=ai-studio) to create one.
27
-
28
-
- Azure OpenAI Deployment with GPT model supporting `chat completion`, for example `gpt-4`.
26
+
- Azure AI Foundry project in the same supported [regions](../../concepts/evaluation-evaluators/risk-safety-evaluators.md#azure-ai-foundry-project-configuration-and-region-support) as risk and safety evaluators (preview). If you don't have an existing project, create one by following the guide [How to create Azure AI Foundry project](../create-projects.md?tabs=ai-studio).
27
+
- Azure OpenAI Deployment with GPT model supporting `chat completion`. For example, `gpt-4`.
29
28
- Make sure you're first logged into your Azure subscription by running `az login`.
30
29
31
-
If this is your first time running evaluations and logging it to your Azure AI Foundry project, you might need to do a few additional setup steps.
30
+
If this is your first time running evaluations and logging it to your Azure AI Foundry project, you might need to do a few additional steps.
32
31
33
-
1.[Create and connect your storage account](https://github.com/azure-ai-foundry/foundry-samples/blob/main/samples/microsoft/infrastructure-setup/01-connections/connection-storage-account.bicep) to your Azure AI Foundry project at the resource level. This bicep template provisions and connects a storage account to your Foundry project with key authentication.
32
+
1.[Create and connect your storage account](https://github.com/azure-ai-foundry/foundry-samples/blob/main/samples/microsoft/infrastructure-setup/01-connections/connection-storage-account.bicep) to your Azure AI Foundry project at the resource level. The [bicep template](https://github.com/azure-ai-foundry/foundry-samples/blob/main/samples/microsoft/infrastructure-setup/01-connections/connection-storage-account.bicep) provisions and connects a storage account to your Foundry project by using key authentication.
34
33
2. Make sure the connected storage account has access to all projects.
35
-
3. If you connected your storage account with Microsoft Entra ID, make sure to give MSI (Microsoft Identity) permissions for Storage Blob Data Ownerto both your account and Foundry project resource in Azure portal.
34
+
3. If you connected your storage account with Microsoft Entra ID, make sure to give MSI (Microsoft Identity) **Storage Blob Data Owner** permissions to both your account and the Foundry project resource in the Azure portal.
36
35
37
-
### Getting started
36
+
### Get started
38
37
39
-
First, install Azure AI Foundry SDK's project client which runs the evaluations in the cloud
38
+
1. Install the Azure AI Foundry SDK project client that runs the evaluations in the cloud.
40
39
41
40
```python
42
41
uv install azure-ai-projects azure-identity
43
42
```
44
43
45
44
> [!NOTE]
46
-
> For more detailed information, see the [REST API Reference Documentation](/rest/api/aifoundry/aiprojects/evaluations).
47
-
Then, set your environment variables for your Azure AI Foundry resources
45
+
> For more detailed information, see [REST API Reference Documentation](/rest/api/aifoundry/aiprojects/evaluations).
46
+
47
+
1. Set your environment variables for your Azure AI Foundry resources
# Upload a local jsonl file (skip if you already have a Dataset registered)
79
+
# Upload a local JSONL file (skip if you already have a Dataset registered)
81
80
data_id = project_client.datasets.upload_file(
82
81
name=dataset_name,
83
82
version=dataset_version,
84
83
file_path="./evaluate_test_data.jsonl",
85
84
).id
86
85
```
87
86
88
-
To learn more about input data formats for evaluating GenAI applications, see [single-turn data](./evaluate-sdk.md#single-turn-support-for-text), [conversation data](./evaluate-sdk.md#conversation-support-for-text), and [conversation data for images and multi-modalities](./evaluate-sdk.md#conversation-support-for-images-and-multi-modal-text-and-image).
87
+
To learn more about input data formats for evaluating generative AI applications, see [Single-turn data](./evaluate-sdk.md#single-turn-support-for-text), [Conversation data](./evaluate-sdk.md#conversation-support-for-text), and [Conversation data for images and multi-modalities](./evaluate-sdk.md#conversation-support-for-images-and-multi-modal-text-and-image).
89
88
90
-
To learn more about input data formats for evaluating agents, see [evaluating Azure AI agents](./agent-evaluate-sdk.md#evaluate-azure-ai-agents) and [evaluating other agents](./agent-evaluate-sdk.md#evaluating-other-agents).
89
+
To learn more about input data formats for evaluating agents, see [Evaluating Azure AI agents](./agent-evaluate-sdk.md#evaluate-azure-ai-agents) and [Evaluating other agents](./agent-evaluate-sdk.md#evaluating-other-agents).
After registering your custom evaluator to your Azure AI project, you can view it in your [Evaluator library](../evaluate-generative-ai-app.md#view-and-manage-the-evaluators-in-the-evaluator-library) under **Evaluation** tab in your Azure AI project.
196
+
After you register your custom evaluator to your Azure AI project, you can view it in your [Evaluator library](../evaluate-generative-ai-app.md#view-and-manage-the-evaluators-in-the-evaluator-library) under the**Evaluation** tab in your Azure AI project.
After logging your custom evaluator to your Azure AI project, you can view it in your [Evaluator library](../evaluate-generative-ai-app.md#view-and-manage-the-evaluators-in-the-evaluator-library) under **Evaluation** tab of your Azure AI project.
244
+
After you log your custom evaluator to your Azure AI project, you can view it in your [Evaluator library](../evaluate-generative-ai-app.md#view-and-manage-the-evaluators-in-the-evaluator-library) under the**Evaluation** tab of your Azure AI project.
246
245
247
246
## Related content
248
247
249
-
-[Evaluate your Generative AI applications locally](./evaluate-sdk.md)
250
-
-[Evaluate your Generative AI applications online](https://aka.ms/GenAIMonitoringDoc)
248
+
-[Evaluate your generative AI applications locally](./evaluate-sdk.md)
249
+
-[Evaluate your generative AI applications online](https://aka.ms/GenAIMonitoringDoc)
251
250
-[Learn more about simulating test datasets for evaluation](./simulator-interaction-data.md)
252
-
-[View your evaluation results in Azure AI project](../../how-to/evaluate-results.md)
253
-
-[Get started building a chat app using the Azure AI Foundry SDK](../../quickstarts/get-started-code.md)
251
+
-[View your evaluation results in an Azure AI project](../../how-to/evaluate-results.md)
252
+
-[Get started building a chat app by using the Azure AI Foundry SDK](../../quickstarts/get-started-code.md)
254
253
-[Get started with evaluation samples](https://aka.ms/aistudio/eval-samples)
0 commit comments