Skip to content

Commit 168ffa6

Browse files
committed
minor update
1 parent 9aa44ae commit 168ffa6

File tree

2 files changed

+7
-7
lines changed

2 files changed

+7
-7
lines changed

articles/ai-studio/how-to/develop/cloud-evaluation.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -14,13 +14,13 @@ ms.reviewer: changliu2
1414
ms.author: lagayhar
1515
author: lgayhardt
1616
---
17-
# Cloud evaluation (Preview): evaluate your Generative AI application remotely on the cloud
17+
# Cloud evaluation (preview): evaluate your Generative AI application remotely on the cloud
1818

1919
[!INCLUDE [feature-preview](../../includes/feature-preview.md)]
2020

21-
While Azure AI Evaluation SDK client supports running evaluations locally on your own machine, you may want to delegate the job remotely to the cloud. For example, after you ran local evaluations on small test data to help assess your generative AI application prototypes, now you move into pre-deployment testing and need run evaluations on a large dataset. Cloud evaluation frees you from managing your local compute infrastructure, and enables you integrate evaluations as tests into your CI/CD pipelines. After deployment, you may want to [continuously evaluate](https://aka.ms/GenAIMonitoringDoc) your applications for post-deployment monitoring.
21+
While Azure AI Evaluation SDK client supports running evaluations locally on your own machine, you might want to delegate the job remotely to the cloud. For example, after you ran local evaluations on small test data to help assess your generative AI application prototypes, now you move into pre-deployment testing and need run evaluations on a large dataset. Cloud evaluation frees you from managing your local compute infrastructure, and enables you to integrate evaluations as tests into your CI/CD pipelines. After deployment, you may want to [continuously evaluate](https://aka.ms/GenAIMonitoringDoc) your applications for post-deployment monitoring.
2222

23-
In this article, you learn how to run cloud evaluation in pre-deployment testing on a test dataset. Using the Azure AI Projects SDK, you will have evaluation results automatically logged into your Azure AI project for better observability. This feature support all Microsft-curated [built-in evaluators](./evaluate-sdk.md#built-in-evaluators) and your own [custom evaluators](./evaluate-sdk.md#custom-evaluators) which can be located in the [Evaluator library](../evaluate-generative-ai-app.md#view-and-manage-the-evaluators-in-the-evaluator-library) of your project.
23+
In this article, you learn how to run cloud evaluation in pre-deployment testing on a test dataset. Using the Azure AI Projects SDK, you'll have evaluation results automatically logged into your Azure AI project for better observability. This feature supports all Microsft-curated [built-in evaluators](./evaluate-sdk.md#built-in-evaluators) and your own [custom evaluators](./evaluate-sdk.md#custom-evaluators) which can be located in the [Evaluator library](../evaluate-generative-ai-app.md#view-and-manage-the-evaluators-in-the-evaluator-library) and have the same project-scope RBAC.
2424

2525

2626
### Prerequisites
@@ -100,15 +100,15 @@ from azure.ai.evaluation import F1ScoreEvaluator, RelevanceEvaluator, ViolenceEv
100100
print("F1 Score evaluator id:", F1ScoreEvaluator.id)
101101
```
102102
103-
- **From UI**: Follows these steps to fetch evaluator ids after they're registered to your project:
103+
- **From UI**: Follows these steps to fetch evaluator IDs after they're registered to your project:
104104
- Select **Evaluation** tab in your Azure AI project;
105105
- Select Evaluator library;
106106
- Select your evaluators of choice by comparing the descriptions;
107107
- Copy its "Asset ID" which will be your evaluator id, for example, `azureml://registries/azureml/models/Groundedness-Evaluator/versions/1`.
108108

109109
#### Specifying custom evaluators
110110

111-
- For code-based custom evaluators, register them to your Azure AI project and fetch the evaluator ids as in this example:
111+
- For code-based custom evaluators, register them to your Azure AI project and fetch the evaluator IDs as in this example:
112112

113113
```python
114114
from azure.ai.ml import MLClient
@@ -199,7 +199,7 @@ After logging your custom evaluator to your Azure AI project, you can view it in
199199
200200
### Cloud evaluation (preview) with Azure AI Projects SDK
201201
202-
Given the steps above, you can now submit a cloud evaluation with Azure AI Projects SDK via a Python API. See the following example specifying an NLP evaluator (F1 score), an AI-assisted quality evaluator (Relevance), a safety evaluator (Violence) and a custom evaluator (Friendliness) with their [evaluator ids](#specifying-evaluators-from-evaluator-library):
202+
You can now submit a cloud evaluation with Azure AI Projects SDK via a Python API. See the following example specifying an NLP evaluator (F1 score), AI-assisted quality and safety evaluator (Relevance and Violence), and a custom evaluator (Friendliness) with their [evaluator IDs](#specifying-evaluators-from-evaluator-library):
203203
204204
```python
205205
import os, time

articles/ai-studio/how-to/develop/evaluate-sdk.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -328,7 +328,7 @@ For conversation outputs, per-turn results are stored in a list and the overall
328328
> [!NOTE]
329329
> We strongly recommend users to migrate their code to use the key without prefixes (for example, `groundedness.groundedness`) to allow your code to support more evaluator models.
330330
331-
### Risk and safety evaluators (Preview)
331+
### Risk and safety evaluators (preview)
332332

333333
When you use AI-assisted risk and safety metrics, a GPT model isn't required. Instead of `model_config`, provide your `azure_ai_project` information. This accesses the Azure AI project safety evaluations back-end service, which provisions a GPT model specific to harms evaluation that can generate content risk severity scores and reasoning to enable the safety evaluators.
334334

0 commit comments

Comments
 (0)