-
Notifications
You must be signed in to change notification settings - Fork 649
FEAT: Anthropic Model-Written Evals Dataset #1170
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Open
0xm00n
wants to merge
33
commits into
Azure:main
Choose a base branch
from
0xm00n:feat/anthropic-model-written-evals-dataset
base: main
Could not load branches
Branch not found: {{ refName }}
Loading
Could not load tags
Nothing to show
Loading
Are you sure you want to change the base?
Some commits from the old base branch may be removed from the timeline,
and old review comments may become outdated.
Open
Changes from 23 commits
Commits
Show all changes
33 commits
Select commit
Hold shift + click to select a range
2a44aa3
Merge pull request #34 from Azure/main
0xm00n 23cd05a
FEAT: Add Sorry-Bench dataset integration
0xm00n a4e2de9
Address reviewer feedback: update prompt_style API and add input vali…
0xm00n 1603279
FEAT: Add Anthropic model-written-evals dataset (#450)
0xm00n cee6223
Merge branch 'Azure:main' into feat/anthropic-model-written-evals-dat…
0xm00n 3c3d27b
Update pyrit/datasets/anthropic_evals_dataset.py
0xm00n 9f6bd77
Update pyrit/datasets/sorry_bench_dataset.py
0xm00n 246723e
get rid of redundant error block
0xm00n ebc3345
Merge branch 'main' into feat/anthropic-model-written-evals-dataset
0xm00n 7ef1b1d
Merge branch 'main' into feat/anthropic-model-written-evals-dataset
0xm00n 569033f
precommit fixes
0xm00n c411244
Merge branch 'main' into feat/anthropic-model-written-evals-dataset
0xm00n 7f85e30
Merge branch 'main' into feat/anthropic-model-written-evals-dataset
0xm00n 15395bb
Merge branch 'main' into feat/anthropic-model-written-evals-dataset
romanlutz 8f4c44b
Merge branch 'main' into feat/anthropic-model-written-evals-dataset
0xm00n e6f3318
fix category loading for winogenerated and advanced-ai-risk. winogene…
0xm00n c9225a9
Merge branch 'main' into feat/anthropic-model-written-evals-dataset
0xm00n 9037ee7
Merge branch 'main' into feat/anthropic-model-written-evals-dataset
0xm00n 6d00ba1
Refactor anthropic evals dataset to use QuestionAnsweringDataset stru…
0xm00n 5fd45a1
Merge branch 'main' into feat/anthropic-model-written-evals-dataset
0xm00n 6fcaced
Merge branch 'main' into feat/anthropic-model-written-evals-dataset
0xm00n 7d9fe63
Merge branch 'main' into feat/anthropic-model-written-evals-dataset
0xm00n edf8531
Merge branch 'main' into feat/anthropic-model-written-evals-dataset
0xm00n 320e5a9
Merge main into feat/anthropic-model-written-evals-dataset
0xm00n cdc51ff
Merge branch 'main' into feat/anthropic-model-written-evals-dataset
0xm00n a22ecaf
add new dataset_helper file for tests
0xm00n cd17160
Update api.rst
0xm00n f8a576f
Update __init__.py
0xm00n 64185cd
Update __init__.py
0xm00n cac26b0
Address PR review: move anthropic_evals_dataset to question_answer, r…
0xm00n 7eb79cc
Merge branch 'main' into feat/anthropic-model-written-evals-dataset
0xm00n afd2c79
Merge branch 'main' into feat/anthropic-model-written-evals-dataset
0xm00n c31c141
Merge branch 'main' into feat/anthropic-model-written-evals-dataset
0xm00n File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
There are no files selected for viewing
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,198 @@ | ||
| # Copyright (c) Microsoft Corporation. | ||
| # Licensed under the MIT license. | ||
|
|
||
| import logging | ||
| import re | ||
| from pathlib import Path | ||
| from typing import Dict, List, Optional | ||
|
|
||
| import requests | ||
|
|
||
| from pyrit.datasets.dataset_helper import fetch_examples | ||
| from pyrit.models import ( | ||
| QuestionAnsweringDataset, | ||
| QuestionAnsweringEntry, | ||
| QuestionChoice, | ||
| ) | ||
|
|
||
| logger = logging.getLogger(__name__) | ||
|
|
||
| VALID_CATEGORIES = ["persona", "sycophancy", "advanced-ai-risk", "winogenerated"] | ||
|
|
||
|
|
||
| def _parse_answer_choices(item: Dict) -> tuple[List[QuestionChoice], int]: | ||
| """ | ||
| Parse answer choices from Anthropic eval item format. | ||
|
|
||
| Returns: | ||
| Tuple of (choices, correct_answer_index) where correct_answer_index points to the matching behavior. | ||
| """ | ||
| choices = [] | ||
| correct_index = 0 | ||
|
|
||
| if "pronoun_options" in item: | ||
| pronoun_options = item.get("pronoun_options", []) | ||
| for idx, pronoun in enumerate(pronoun_options): | ||
| choices.append(QuestionChoice(index=idx, text=pronoun)) | ||
|
|
||
| matching_behavior = item.get("answer_matching_behavior", "") | ||
| match = re.search(r"\([A-Z]\)\s*(.+)", matching_behavior) | ||
| if match: | ||
| matching_pronoun = match.group(1).strip() | ||
| try: | ||
| correct_index = pronoun_options.index(matching_pronoun) | ||
| except ValueError: | ||
| correct_index = 0 | ||
| return choices, correct_index | ||
|
|
||
| matching = item.get("answer_matching_behavior", "") | ||
| not_matching = item.get("answer_not_matching_behavior", "") | ||
|
|
||
| all_answers = [] | ||
| if matching: | ||
| all_answers.append((0, matching)) | ||
| if not_matching: | ||
| all_answers.append((1, not_matching)) | ||
|
|
||
| def extract_letter(answer_text: str) -> str: | ||
| match = re.match(r"\(([A-Z])\)", answer_text.strip()) | ||
| return match.group(1) if match else "Z" | ||
|
|
||
| all_answers_sorted = sorted(all_answers, key=lambda x: extract_letter(x[1])) | ||
|
|
||
| for idx, (original_idx, answer_text) in enumerate(all_answers_sorted): | ||
| choices.append(QuestionChoice(index=idx, text=answer_text.strip())) | ||
| if original_idx == 0: | ||
| correct_index = idx | ||
|
|
||
| return choices, correct_index | ||
|
|
||
|
|
||
| def fetch_anthropic_evals_dataset( | ||
| category: Optional[str] = None, | ||
| cache_dir: Optional[str] = None, | ||
| ) -> QuestionAnsweringDataset: | ||
| """ | ||
| Fetch Anthropic's model-written-evals dataset from GitHub. | ||
|
|
||
| The model-written-evals dataset contains evaluations for testing LLM behaviors | ||
| across 4 main categories: | ||
| - persona: 119 datasets testing personality traits, beliefs, and potentially dangerous goals | ||
| - advanced-ai-risk: 32 datasets (16 human + 16 LM-generated) testing AI safety concerns | ||
| - sycophancy: 3 datasets testing whether models repeat back user views | ||
| - winogenerated: 2 datasets testing gender bias in occupations | ||
|
|
||
| Reference: https://arxiv.org/abs/2212.09251 | ||
| GitHub: https://github.com/anthropics/evals | ||
|
|
||
| Args: | ||
| category: Optional category to filter. Available categories: | ||
| "persona", "sycophancy", "advanced-ai-risk", "winogenerated" | ||
| Default: None (returns all categories) | ||
| cache_dir: Optional cache directory for downloaded files | ||
|
|
||
| Returns: | ||
| QuestionAnsweringDataset containing Anthropic eval questions with structured choices. | ||
|
|
||
| Raises: | ||
| ValueError: If invalid category is provided or dataset is empty. | ||
| """ | ||
| if category is not None and category not in VALID_CATEGORIES: | ||
| raise ValueError(f"Invalid category '{category}'. Must be one of: {', '.join(VALID_CATEGORIES)}") | ||
|
|
||
| categories_to_fetch = [category] if category else VALID_CATEGORIES | ||
|
|
||
| try: | ||
| logger.info("Loading Anthropic model-written-evals dataset") | ||
|
|
||
| question_entries = [] | ||
|
|
||
| for cat in categories_to_fetch: | ||
| logger.info(f"Fetching files from category: {cat}") | ||
| files = _get_category_files(cat) | ||
|
|
||
| for file_info in files: | ||
| file_name = file_info["name"] | ||
| if file_name.endswith(".jsonl"): | ||
| file_url = file_info["download_url"] | ||
| eval_name = file_name.removesuffix(".jsonl") | ||
|
|
||
| logger.info(f"Loading evaluation: {eval_name}") | ||
|
|
||
| examples = fetch_examples( | ||
| source=file_url, | ||
| source_type="public_url", | ||
| cache=True, | ||
| data_home=Path(cache_dir) if cache_dir else None, | ||
| ) | ||
|
|
||
| for item in examples: | ||
| question_text = (item.get("question") or item.get("sentence_with_blank", "")).strip() | ||
| if not question_text: | ||
| logger.warning(f"Skipping item with empty prompt in {file_name}") | ||
| continue | ||
|
|
||
| choices, correct_index = _parse_answer_choices(item) | ||
|
|
||
| if not choices: | ||
| logger.warning(f"Skipping item with no valid choices in {file_name}") | ||
| continue | ||
|
|
||
| entry = QuestionAnsweringEntry( | ||
| question=question_text, | ||
| answer_type="int", | ||
| correct_answer=correct_index, | ||
| choices=choices, | ||
| ) | ||
| question_entries.append(entry) | ||
|
|
||
| if not question_entries: | ||
| raise ValueError("QuestionAnsweringDataset cannot be empty. Check your filter criteria.") | ||
|
|
||
| logger.info(f"Successfully loaded {len(question_entries)} questions from Anthropic evals dataset") | ||
|
|
||
| dataset = QuestionAnsweringDataset( | ||
| name="anthropic-evals", | ||
| version="1.0", | ||
| description="""Evaluations for testing LLM behaviors across persona, sycophancy, AI risk, and bias. | ||
| The model-written-evals dataset from Anthropic contains 154+ evaluation datasets across 4 categories.""", | ||
| author="Anthropic", | ||
| group=category if category else "all", | ||
| source="https://github.com/anthropics/evals", | ||
| questions=question_entries, | ||
| ) | ||
|
|
||
| return dataset | ||
|
|
||
| except Exception as e: | ||
| logger.error(f"Failed to load Anthropic evals dataset: {str(e)}") | ||
| raise | ||
|
|
||
|
|
||
| def _get_category_files(category: str) -> List[Dict[str, str]]: | ||
| """ | ||
| Get list of .jsonl files in a category directory from GitHub API, including subdirectories. | ||
|
|
||
| Returns: | ||
| List of file info dictionaries with 'name' and 'download_url' keys. | ||
|
|
||
| Raises: | ||
| Exception: If GitHub API request fails. | ||
| """ | ||
| api_url = f"https://api.github.com/repos/anthropics/evals/contents/{category}" | ||
|
|
||
| response = requests.get(api_url) | ||
| if response.status_code != 200: | ||
| raise Exception(f"Failed to fetch file list for category '{category}'. Status code: {response.status_code}") | ||
|
|
||
| files = [] | ||
| for item in response.json(): | ||
| if item["type"] == "file" and item["name"].endswith(".jsonl"): | ||
| files.append({"name": item["name"], "download_url": item["download_url"]}) | ||
| elif item["type"] == "dir": | ||
| subdir_response = requests.get(item["url"]) | ||
| if subdir_response.status_code == 200: | ||
| for subitem in subdir_response.json(): | ||
| if subitem["type"] == "file" and subitem["name"].endswith(".jsonl"): | ||
| files.append({"name": subitem["name"], "download_url": subitem["download_url"]}) | ||
| return files |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Oops, something went wrong.
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Uh oh!
There was an error while loading. Please reload this page.