You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
> Evaluation with the prompt flow SDK has been retired and replaced with Azure AI Evaluation SDK client library for Python. For more information about input data requirements, see the [API Reference Documentation](https://aka.ms/azureaieval-python-ref).
22
+
> For more information about input data requirements, see the [API Reference Documentation](https://aka.ms/azureaieval-python-ref).
23
23
24
24
To thoroughly assess the performance of your generative AI application when applied to a substantial dataset, you can evaluate a Generative AI application in your development environment with the Azure AI evaluation SDK. Given either a test dataset or a target, your generative AI application generations are quantitatively measured with both mathematical based metrics and AI-assisted quality and safety evaluators. Built-in or custom evaluators can provide you with comprehensive insights into the application's capabilities and limitations.
25
25
@@ -38,6 +38,7 @@ pip install azure-ai-evaluation
38
38
Built-in evaluators support the following application scenarios:
39
39
40
40
-**Query and response**: This scenario is designed for applications that involve sending in queries and generating responses, usually single-turn.
41
+
-**Conversation**: This scenario is designed for applications that involve sending in queries and generating responses in a multi-turn exchange.
41
42
-**Retrieval augmented generation**: This scenario is suitable for applications where the model engages in generation using a retrieval-augmented approach to extract information from your provided documents and generate detailed responses, usually multi-turn.
42
43
43
44
For more in-depth information on each evaluator definition and how it's calculated, see [Evaluation and monitoring metrics for generative AI](../../concepts/evaluation-metrics-built-in.md).
@@ -46,7 +47,7 @@ For more in-depth information on each evaluator definition and how it's calculat
0 commit comments