Skip to content

Add evalview-agent-testing — regression testing for AI agents#153

Open
hidai25 wants to merge 1 commit intoBehiSecc:mainfrom
hidai25:feat/add-evalview
Open

Add evalview-agent-testing — regression testing for AI agents#153
hidai25 wants to merge 1 commit intoBehiSecc:mainfrom
hidai25:feat/add-evalview

Conversation

@hidai25
Copy link

@hidai25 hidai25 commented Mar 23, 2026

Adds EvalView to the development tools section.

EvalView provides regression testing for AI agents — snapshot behavior, detect when tool calls or output quality drift, and block broken agents before production.

Features:

  • Golden baseline diffing
  • Python API (gate()) for autonomous loops
  • MCP server (8 tools) for Claude Code
  • CI/CD with PR comments
  • Multi-turn test support

Published on PyPI (pip install evalview), Apache 2.0 license.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant