-
-
Notifications
You must be signed in to change notification settings - Fork 1.1k
Description
Display Name
sniffbench
Category
Tooling
Sub-Category
General
Primary Link
https://github.com/AnswerLayer/sniffbench
Secondary Link
https://www.npmjs.com/package/sniffbench
Author Name
AnswerLayer
Author Link
https://github.com/AnswerLayer
License
MIT
Other License
No response
Description
A benchmark suite for evaluating Claude Code configurations. Run A/B tests comparing different CLAUDE.md files, model settings, or permission configurations using real issues from your repos. Track metrics like token usage, cost, and task completion across test runs.
Validate Claims
Install via npm install -g sniffbench, then run sniff interview to see Claude Code answer comprehension questions about your codebase. Compare results across different configurations to measure improvements.
Additional Comments
sniffbench helps teams make data-driven decisions about their Claude Code setup by providing reproducible benchmarks rather than relying on subjective impressions.
Recommendation Checklist
- I have checked that this resource hasn't already been submitted
- My resource provides genuine value to Claude Code users, and any risks are clearly stated
- All provided links are working and publicly accessible
- I am submitting only ONE resource in this issue
- I understand that low-quality or duplicate submissions may be rejected