Skip to content

Conversation

@dependabot
Copy link
Contributor

@dependabot dependabot bot commented on behalf of github Nov 10, 2025

Bumps azure-ai-evaluation from 1.8.0 to 1.13.4.

Release notes

Sourced from azure-ai-evaluation's releases.

azure-ai-evaluation_1.13.4

1.13.4 (2025-11-10)

Bugs Fixed

  • Handle input data for evaluation result when evaluators.

azure-ai-evaluation_1.13.3

1.13.3 (2025-11-08)

Other Changes

  • Added scenario property to red team evaluation request to align scores with red team concepts of attack success.

azure-ai-evaluation_1.13.2

1.13.2 (2025-11-07)

Bugs Fixed

  • Added App Insights redaction for agent safety run telemetry so adversarial prompts are not stored in collected logs.

azure-ai-evaluation_1.13.1

1.13.1 (2025-11-05)

Features Added

  • Improved RedTeam coverage across risk sub-categories to ensure comprehensive security testing
  • Made RedTeam's AttackStrategy.Tense seed prompts dynamic to allow use of this strategy with additional risk categories
  • Refactors error handling and result semantics in the RedTeam evaluation system to improve clarity and align with Attack Success Rate (ASR) conventions (passed=False means attack success)

Bugs Fixed

  • Fixed RedTeam evaluation error related to context handling for context-dependent risk categories
  • Fixed RedTeam prompt application for model targets during Indirect Jailbreak XPIA (Cross-Platform Indirect Attack)

azure-ai-evaluation_1.13.0

1.13.0 (2025-10-30)

Features Added

  • Updated IndirectAttack risk category for RedTeam to IndirectJailbreak to better reflect its purpose. This change allows users to apply cross-domain prompt injection (XPIA) attack strategies across all risk categories, enabling more comprehensive security testing of AI systems against indirect prompt injection attacks during red teaming.
  • Added TaskAdherence, SensitiveDataLeakage, and ProhibitedActions as cloud-only agent safety risk categories for red teaming.
  • Updated all evaluators' output to be of the following schema:
    • gpt_{evaluator_name}, {evaluator_name}: float score,
    • {evaluator_name}_result: pass/fail based on threshold,
    • {evaluator_name}_reason, {evaluator_name}_threshold
    • {evaluator_name}_prompt_tokens, {evaluator_name}_completion_tokens, {evaluator_name}_total_tokens, {evaluator_name}_finish_reason
    • {evaluator_name}_model: model used for evaluation
    • {evaluator_name}_sample_input, {evaluator_name}_sample_output: input and output used for evaluation

... (truncated)

Commits
  • d35f12e Handle the input data for pure string_check evaluation in result converter (#...
  • d0c3eea Rename Tool Success to Tool Call Success (#43810)
  • bb35e87 updated (#43898)
  • 8ac8b3e [EvaluationResult Convert]Counts only for primary metrics when multiple metri...
  • 484d5d0 fix dependency issue in agentserver (#43891)
  • 32d6342 [ai.agentserver] add azure-ai-agentserver-agentframework and azure-ai-agentse...
  • c75311a Allow parameterization of python version during analyze (#43888)
  • 90a9226 Added Autoscale Throughput Samples (#43619)
  • 09f0066 [Storage] Added Support for UseDevelopmentStorage=true; for Connection Stri...
  • b7dc156 [Storage] Changelogs for Bug Fixes (#43808)
  • Additional commits viewable in compare view

Dependabot compatibility score

Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting @dependabot rebase.


Dependabot commands and options

You can trigger Dependabot actions by commenting on this PR:

  • @dependabot rebase will rebase this PR
  • @dependabot recreate will recreate this PR, overwriting any edits that have been made to it
  • @dependabot merge will merge this PR after your CI passes on it
  • @dependabot squash and merge will squash and merge this PR after your CI passes on it
  • @dependabot cancel merge will cancel a previously requested merge and block automerging
  • @dependabot reopen will reopen this PR if it is closed
  • @dependabot close will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually
  • @dependabot show <dependency name> ignore conditions will show all of the ignore conditions of the specified dependency
  • @dependabot ignore this major version will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself)
  • @dependabot ignore this minor version will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself)
  • @dependabot ignore this dependency will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)

Bumps [azure-ai-evaluation](https://github.com/Azure/azure-sdk-for-python) from 1.8.0 to 1.13.4.
- [Release notes](https://github.com/Azure/azure-sdk-for-python/releases)
- [Changelog](https://github.com/Azure/azure-sdk-for-python/blob/main/doc/esrp_release.md)
- [Commits](Azure/azure-sdk-for-python@azure-ai-evaluation_1.8.0...azure-ai-evaluation_1.13.4)

---
updated-dependencies:
- dependency-name: azure-ai-evaluation
  dependency-version: 1.13.4
  dependency-type: direct:production
  update-type: version-update:semver-minor
...

Signed-off-by: dependabot[bot] <[email protected]>
@dependabot dependabot bot added dependencies Pull requests that update a dependency file python Pull requests that update Python code labels Nov 10, 2025
@dependabot @github
Copy link
Contributor Author

dependabot bot commented on behalf of github Nov 24, 2025

Superseded by #156.

@dependabot dependabot bot closed this Nov 24, 2025
@dependabot dependabot bot deleted the dependabot/pip/azure-ai-evaluation-1.13.4 branch November 24, 2025 08:15
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

dependencies Pull requests that update a dependency file python Pull requests that update Python code

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant