__ __ __
____ ______/ /_/ /_ ___ ____ _____/ /_
/ __ `/ ___/ __/ __ \/ _ \/ __ \/ ___/ __ \
/ /_/ / /__/ /_/ /_/ / __/ / / / /__/ / / /
\__,_/\___/\__/_.___/\___/_/ /_/\___/_/ /_/
actbench is a extensible framework designed to evaluate the performance and capabilities of web automation agents and LAM systems.
actbench requires Python 3.12 or higher. We recommend using pipx for a clean, isolated installation:
pipx install actbenchBefore running benchmarks, you need to set API keys for the agents you want to use.
actbench set-key --agent raccoonaiYou can list the supported agents and check which API keys are stored:
actbench agents listactbench provides a built-in dataset of web automation tasks, crafted by merging and refining tasks from the webarena and webvoyager datasets.
Duplicate tasks have been stripped out, and the queries have been refreshed to align with the most recent information.
If you want to explore how the tasks have been modified, you can trace their IDs back to the original datasets for a side-by-side comparison.
To see all the tasks currently available, just run this command:
actbench tasks listThe run command is the heart of actbench. It allows you to execute tasks against specified agents.
actbench run --agent raccoonai --task 256 --task 424This command runs tasks with IDs 256 and 424 using the raccoonai agent.
actbench run --agent raccoonai --all-tasksThis runs all available tasks using the raccoonai agent.
actbench run --agent raccoonai --random 5This runs a random sample of 5 tasks using the raccoonai agent.
actbench run --all-agents --all-tasksThis runs all tasks with all configured agents (for which API keys are stored).
actbench run --agent raccoonai --all-tasks --parallel 4This runs all tasks using the raccoonai agent, executing up to 4 tasks concurrently.
actbench run --agent raccoonai --all-tasks --rate-limit 0.5This adds a 0.5-second delay between task submissions.
actbench run --agent raccoonai --all-tasks --no-scoringThis disables the LLM powered scoring, and gives all tasks a score of -1.
You can combine these options for more complex benchmark configurations:
actbench run --agent raccoonai --agent anotheragent --task 1 --task 2 --random 3 --parallel 2 --rate-limit 0.2This command runs tasks 1 and 2, plus 3 random tasks, using both raccoonai and anotheragent (assuming API keys are set), with a parallelism of 2 and a rate limit of 0.2 seconds.
The results command group allows you to manage and view benchmark results.
actbench results listYou can filter results by agent or run ID:
actbench results list --agent raccoonai
actbench results list --run-id <run_id>You can export results to JSON or CSV files:
actbench results export --format json --output results.json
actbench results export --format csv --output results.csv --agent raccoonaiHere's a complete table detailing the actbench CLI commands, their flags (options) and explanations:
| Command | Flag(s) / Option(s) | Explanation |
|---|---|---|
actbench run |
--task / -t |
Specifies one or more task IDs to run. Can be used multiple times. If omitted, other task selection flags (--random, --all-tasks) must be used. |
--agent / -a |
Specifies one or more agents to use. Can be used multiple times. If omitted, --all-agents must be used. |
|
--random / -r |
Runs a specified number of random tasks. Takes an integer argument (e.g., --random 5). |
|
--all-tasks |
Runs all available tasks. | |
--all-agents |
Runs with all configured agents (for which API keys have been set). | |
--parallel / -p |
Sets the number of tasks to run concurrently. Takes an integer argument (e.g., --parallel 4). Defaults to 1 (no parallelism). |
|
--rate-limit / -l |
Sets the delay (in seconds) between task submissions. Takes a float argument (e.g., --rate-limit 0.5). Defaults to 0.1. |
|
--no-scoring / -ns |
Disables LLM-based scoring. Results will have a score of -1. | |
actbench tasks list |
None | Lists all available tasks in the dataset, showing their ID, query, URL, complexity, and whether they require login. |
actbench set-key |
--agent / -a |
Sets the API key for a specified agent. Prompts the user to enter the key securely. Example: actbench set-key --agent raccoonai |
actbench agents list |
None | Lists all supported agents, and shows which agents have API Keys stored. |
actbench results list |
--agent / -a |
Filters the results to show only those for a specific agent. |
--run-id / -r |
Filters the results to show only those for a specific run ID. | |
actbench results export |
--agent / -a |
Filters the results to be exported to a specific agent. |
--run-id / -r |
Filters the results to be exported for a specific run ID. | |
--format / -f |
Specifies the export format. Must be one of json or csv. Defaults to json. |
|
--output / -o |
Specifies the output file path. Required. | |
actbench |
None | Prints the help message for the CLI. |
actbench --version |
None | Prints the actbench version number. |
- Create a new client class: Create a new Python file in the
actbench/clients/directory (e.g.,my_agent.py). - Implement the
BaseClientinterface: Your class should inherit fromactbench.clients.BaseClientand implement theset_api_key()andrun()methods. - Register your client: Add your client class to the
_CLIENT_REGISTRYinactbench/clients/__init__.py.
- Create a new dataset class: Create a new Python file in the
actbench/datasets/directory (e.g.,my_dataset.py). - Implement the
BaseDatasetinterface: Your class should inherit fromactbench.datasets.BaseDatasetand implement theload_task_data(),get_all_task_ids(), andget_all_tasks()methods. - Provide your dataset file: Place your dataset file (e.g.,
my_dataset.jsonl) in thesrc/actbench/dataset/directory. - Update
_DATASET_INSTANCE: If you want to use this dataset by default, update the_DATASET_INSTANCEvariable insrc/actbench/datasets/__init__.py.
You can customize the evaluation process by modifying the Evaluator class in actbench/executor/evaluator.py or by creating a new evaluator and integrating it into the TaskExecutor.
Contributions are welcome! Please follow these simple guidelines:
- Fork the repository.
- Create a new branch for your feature or bug fix.
- Write clear and concise code with appropriate comments.
- Submit a pull request.