Skip to content

Commit 3083053

Browse files
authored
Update eval tutorial order (#46343)
1 parent decd4c8 commit 3083053

File tree

4 files changed

+16
-14
lines changed

4 files changed

+16
-14
lines changed

docs/ai/quickstarts/evaluate-ai-response.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ ms.topic: quickstart
66
ms.custom: devx-track-dotnet, devx-track-dotnet-ai
77
---
88

9-
# Evaluate the quality of a model's response
9+
# Quickstart: Evaluate response quality
1010

1111
In this quickstart, you create an MSTest app to evaluate the quality of a chat response from an OpenAI model. The test app uses the [Microsoft.Extensions.AI.Evaluation](https://www.nuget.org/packages/Microsoft.Extensions.AI.Evaluation) libraries.
1212

docs/ai/toc.yml

Lines changed: 8 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -81,12 +81,14 @@ items:
8181
items:
8282
- name: The Microsoft.Extensions.AI.Evaluation libraries
8383
href: conceptual/evaluation-libraries.md
84-
- name: "Quickstart: Evaluate the quality of a response"
85-
href: quickstarts/evaluate-ai-response.md
86-
- name: "Tutorial: Evaluate the safety of a response"
87-
href: tutorials/evaluate-safety.md
88-
- name: "Tutorial: Evaluate a response with caching and reporting"
89-
href: tutorials/evaluate-with-reporting.md
84+
- name: Tutorials
85+
items:
86+
- name: "Quickstart: Evaluate the quality of a response"
87+
href: quickstarts/evaluate-ai-response.md
88+
- name: "Evaluate response quality with caching and reporting"
89+
href: tutorials/evaluate-with-reporting.md
90+
- name: "Evaluate response safety with caching and reporting"
91+
href: tutorials/evaluate-safety.md
9092
- name: Resources
9193
items:
9294
- name: API reference

docs/ai/tutorials/evaluate-safety.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,12 @@
11
---
2-
title: Tutorial - Evaluate the content safety of a model's response
3-
description: Create an MSTest app that evaluates the content safety of a model's response using the evaluators in the Microsoft.Extensions.AI.Evaluation.Safety package.
2+
title: Tutorial - Evaluate response safety with caching and reporting
3+
description: Create an MSTest app that evaluates the content safety of a model's response using the evaluators in the Microsoft.Extensions.AI.Evaluation.Safety package and with caching and reporting.
44
ms.date: 05/12/2025
55
ms.topic: tutorial
66
ms.custom: devx-track-dotnet-ai
77
---
88

9-
# Tutorial: Evaluate the content safety of a model's response
9+
# Tutorial: Evaluate response safety with caching and reporting
1010

1111
In this tutorial, you create an MSTest app to evaluate the *content safety* of a response from an OpenAI model. Safety evaluators check for presence of harmful, inappropriate, or unsafe content in a response. The test app uses the safety evaluators from the [Microsoft.Extensions.AI.Evaluation.Safety](https://www.nuget.org/packages/Microsoft.Extensions.AI.Evaluation.Safety) package to perform the evaluations. These safety evaluators use the [Azure AI Foundry](/azure/ai-foundry/) Evaluation service to perform evaluations.
1212

docs/ai/tutorials/evaluate-with-reporting.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,14 +1,14 @@
11
---
2-
title: Tutorial - Evaluate a model's response
3-
description: Create an MSTest app and add a custom evaluator to evaluate the AI chat response of a language model, and learn how to use the caching and reporting features of Microsoft.Extensions.AI.Evaluation.
2+
title: Tutorial - Evaluate response quality with caching and reporting
3+
description: Create an MSTest app to evaluate the response quality of a language model, add a custom evaluator, and learn how to use the caching and reporting features of Microsoft.Extensions.AI.Evaluation.
44
ms.date: 05/09/2025
55
ms.topic: tutorial
66
ms.custom: devx-track-dotnet-ai
77
---
88

9-
# Tutorial: Evaluate a model's response with response caching and reporting
9+
# Tutorial: Evaluate response quality with caching and reporting
1010

11-
In this tutorial, you create an MSTest app to evaluate the chat response of an OpenAI model. The test app uses the [Microsoft.Extensions.AI.Evaluation](https://www.nuget.org/packages/Microsoft.Extensions.AI.Evaluation) libraries to perform the evaluations, cache the model responses, and create reports. The tutorial uses both built-in and custom evaluators.
11+
In this tutorial, you create an MSTest app to evaluate the chat response of an OpenAI model. The test app uses the [Microsoft.Extensions.AI.Evaluation](https://www.nuget.org/packages/Microsoft.Extensions.AI.Evaluation) libraries to perform the evaluations, cache the model responses, and create reports. The tutorial uses both built-in and custom evaluators. The built-in quality evaluators (from the [Microsoft.Extensions.AI.Evaluation.Quality package](https://www.nuget.org/packages/Microsoft.Extensions.AI.Evaluation.Quality)) use an LLM to perform evaluations; the custom evaluator does not use AI.
1212

1313
## Prerequisites
1414

0 commit comments

Comments
 (0)