Skip to content

Conversation

@YoYoJa
Copy link
Contributor

@YoYoJa YoYoJa commented Nov 7, 2025

Description

Please add an informative description that covers that changes made by the pull request and link all relevant issues.

If an SDK is being regenerated based on a new API spec, a link to the pull request containing these API spec changes should be included above.

All SDK Contribution checklist:

  • The pull request does not introduce [breaking changes]
  • CHANGELOG is updated for new features, bug fixes or other significant changes.
  • I have read the contribution guidelines.

General Guidelines and Best Practices

  • Title of the pull request is clear and informative.
  • There are a small number of commits, each of which have an informative message. This means that previously merged commits do not appear in the history of the PR. For more information on cleaning up the commits in your PR, see this page.

Testing Guidelines

  • Pull request includes test coverage for the included changes.

@YoYoJa YoYoJa requested a review from a team as a code owner November 7, 2025 20:03
Copilot AI review requested due to automatic review settings November 7, 2025 20:03
@github-actions github-actions bot added the Evaluation Issues related to the client library for Azure AI Evaluation label Nov 7, 2025
Copy link
Contributor

Copilot AI left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Pull Request Overview

This PR introduces filtering logic to skip non-primary metrics when calculating AOAI evaluation summaries. The primary metric is defined as the first metric in the list for evaluators that produce multiple metrics.

  • Added a new _is_primary_metric function to determine if a metric is a primary metric
  • Modified _calculate_aoai_evaluation_summary to skip counting non-primary metrics
  • Reordered the rouge_score metrics list to make rouge_f1_score the primary metric instead of rouge_precision

Reviewed Changes

Copilot reviewed 2 out of 2 changed files in this pull request and generated 2 comments.

File Description
sdk/evaluation/azure-ai-evaluation/azure/ai/evaluation/_evaluate/_evaluate.py Added _is_primary_metric function and integrated primary metric filtering into _calculate_aoai_evaluation_summary
sdk/evaluation/azure-ai-evaluation/azure/ai/evaluation/_constants.py Updated documentation for _EvaluatorMetricMapping and reordered rouge_score metrics

@YoYoJa YoYoJa changed the title [EvaluationResult Contert]Counts only for primary metrics when multiple metrics [EvaluationResult Convert]Counts only for primary metrics when multiple metrics Nov 8, 2025
@YoYoJa YoYoJa changed the title [EvaluationResult Convert]Counts only for primary metrics when multiple metrics [EvaluationResult Convert]Counts only for primary metrics when multiple metrics and exclude errored counts for passed/failed Nov 8, 2025
@YoYoJa YoYoJa merged commit 8ac8b3e into main Nov 8, 2025
20 checks passed
@YoYoJa YoYoJa deleted the jessli/UpdateMappings branch November 8, 2025 22:45
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Evaluation Issues related to the client library for Azure AI Evaluation

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants