Skip to content

Commit f13a191

Browse files
authored
Update evaluate-prompts-playground.md
1 parent 433db1d commit f13a191

File tree

1 file changed

+10
-10
lines changed

1 file changed

+10
-10
lines changed

articles/ai-foundry/how-to/evaluate-prompts-playground.md

Lines changed: 10 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
---
2-
title: Manually Evaluate Prompts in Azure AI Foundry Portal Playground
2+
title: Manually Evaluate Prompts in the Azure AI Foundry Portal Playground
33
titleSuffix: Azure AI Foundry
4-
description: Learn how to quickly test and evaluate prompts in Azure AI Foundry portal playground.
4+
description: Learn how to quickly test and evaluate prompts in the Azure AI Foundry portal playground.
55
manager: scottpolly
66
ms.service: azure-ai-foundry
77
ms.custom:
@@ -14,7 +14,7 @@ ms.author: lagayhar
1414
author: lgayhardt
1515
---
1616

17-
# Manually evaluate prompts in Azure AI Foundry portal playground
17+
# Manually evaluate prompts in the Azure AI Foundry portal playground
1818

1919
[!INCLUDE [feature-preview](../includes/feature-preview.md)]
2020

@@ -35,16 +35,16 @@ In this article, you learn to:
3535
## Prerequisites
3636

3737
* A test dataset in one of these formats: CSV or JSON Lines (JSONL). If you don't have a dataset available, you can also manually enter data from the UI.
38-
* A deployment of one of these models: GPT 3.5, GPT 4, or Davinci. To learn more about how to create a deployment, see [Deploy models](./deploy-models-openai.md).
38+
* A deployment of one of these models: GPT-3.5, GPT-4, or Davinci. To learn more about how to create a deployment, see [Deploy models](./deploy-models-openai.md).
3939

4040
> [!NOTE]
4141
> At this time, manual evaluation is only supported for Azure OpenAI models for chat and completion task types.
4242
4343
## Generate your manual evaluation results
4444

45-
From the **Playground**, select the **Manual evaluation** option to begin the process of manually reviewing the model responses based on your test data and prompt. Your prompt is automatically transitioned to your **Manual evaluation** file. You need to add test data to evaluate the prompt against. You can do this step manually by using the text boxes in the **Input** column.
45+
From **Playground**, select the **Manual evaluation** option to begin the process of manually reviewing the model responses based on your test data and prompt. Your prompt is automatically transitioned to your **Manual evaluation** file. You need to add test data to evaluate the prompt against. You can do this step manually by using the text boxes in the **Input** column.
4646

47-
You can also use the **Import Data** feature to select one of the existing datasets in your project, or upload a dataset in CSV or JSONL format. After loading your data, you'll be prompted to map the columns appropriately. After you finish and select **Import**, the data is populated in the appropriate columns.
47+
You can also use the **Import Data** feature to select one of the existing datasets in your project, or upload a dataset in CSV or JSONL format. After you load your data, you're prompted to map the columns appropriately. After you finish and select **Import**, the data is populated in the appropriate columns.
4848

4949
:::image type="content" source="../media/evaluations/prompts/generate-manual-evaluation-results.png" alt-text="Screenshot that shows how to generate manual evaluation results." lightbox= "../media/evaluations/prompts/generate-manual-evaluation-results.png":::
5050

@@ -55,23 +55,23 @@ Now that your data is added, you can select **Run** to populate the output colum
5555

5656
## Rate your model's responses
5757

58-
You can rate the prompt's output by selecting a thumbs up or down for each response. Based on the ratings you provide, you can view these response scores in the at-a-glance summaries.
58+
You can rate the prompt's output by selecting a thumbs up or down for each response. Based on the ratings that you provide, you can view these response scores in the at-a-glance summaries.
5959

6060
:::image type="content" source="../media/evaluations/prompts/rate-results.png" alt-text="Screenshot that shows response scores in the at-a-glance summaries." lightbox= "../media/evaluations/prompts/rate-results.png":::
6161

6262
## Iterate on your prompt and reevaluate
6363

6464
Based on your summary, you might want to make changes to your prompt. You can edit your prompt setup by using the prompt controls mentioned previously. You can update the system message, change the model, edit the parameters, and more.
6565

66-
After making your edits, you can run them all again to update the entire table or run only specific rows again that didn't meet your expectations the first time.
66+
After you make your edits, you can run them all again to update the entire table or run only specific rows again that didn't meet your expectations the first time.
6767

6868
## Save and compare results
6969

70-
After populating your results, you can select **Save results**. By saving your results, you can share the progress with your team or continue your manual evaluation later.
70+
After you populate your results, you can select **Save results**. By saving your results, you can share the progress with your team or continue your manual evaluation later.
7171

7272
:::image type="content" source="../media/evaluations/prompts/save-and-compare-results.png" alt-text="Screenshot of the Save results selection." lightbox= "../media/evaluations/prompts/save-and-compare-results.png":::
7373

74-
You can also compare the thumbs up and down ratings across your manual evaluations. Save them, and then view them in the **Evaluation** tab under **Manual evaluation**.
74+
You can also compare the thumbs up and down ratings across your manual evaluations. Save them, and then view them on the **Evaluation** tab under **Manual evaluation**.
7575

7676
## Related content
7777

0 commit comments

Comments
 (0)