You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
chore: add AGENTS.md and make skills runtime-agnostic (#160)
## Summary
This PR prepares the repository for development with Codex in addition
to Claude Code by introducing Codex-facing entrypoints and making
existing skill instructions runtime-agnostic.
### What changed
1. Added agent compatibility symlinks
- Added `AGENTS.md` as a symlink to `CLAUDE.md`
- Added `.agents/skills` as a symlink to `.claude/skills`
2. Updated skill docs to be runtime-agnostic
- `.claude/skills/golden-test-reviewer/SKILL.md`
- Replaced Claude-specific interaction/tool wording (e.g.
AskUserQuestion, Task tool)
- Clarified isolated subtask execution model and parallelization
requirements
- Added explicit guidance to defer cross-case synthesis to aggregation
step
- `.claude/skills/test-pruning-advisor/SKILL.md`
- Updated references from `CLAUDE.md` to `AGENTS.md / CLAUDE.md`
- Replaced Claude-specific interaction/tool wording with
runtime-agnostic guidance
- Clarified isolated subtasks, minimal context passing, and aggregation
responsibilities
3. Simplified base guidance document
- `CLAUDE.md`
- Removed the Kiro-specific "Development Workflow / AI-DLC and
Spec-Driven Development" section to keep project instructions
tool-agnostic and focused on repository conventions
## Notes
- Documentation/configuration only (symlinks + markdown updates)
- No production code or test behavior changes
Copy file name to clipboardExpand all lines: .claude/skills/golden-test-reviewer/SKILL.md
+25-15Lines changed: 25 additions & 15 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -15,7 +15,7 @@ Systematically evaluate golden test cases under `packages/cli/src/gen-orchestrat
15
15
16
16
### Step 1: Select Review Target
17
17
18
-
When the review target is not explicitly specified, use AskUserQuestion to present the following options:
18
+
When the review target is not explicitly specified, ask the user (using the runtime's available interaction mechanism) to choose from the following options:
19
19
20
20
1.**All**: Review all test cases
21
21
2.**Category**: Review test cases in a specific category only
@@ -110,9 +110,13 @@ This analysis will be included in the final report (Step 3).
110
110
111
111
### Step 2: Launch Subagents for Parallel Evaluation
112
112
113
-
For each test case to be reviewed, launch a subagentusing the Task tool with `subagent_type: "general-purpose"`.
113
+
For each test case to be reviewed, create a separate subtask with an isolated context (subagent/subtask model) using the runtime's available execution mechanism.
114
114
115
-
**IMPORTANT**: Launch all subagents in a single response with multiple Task tool calls to enable parallel execution.
115
+
**IMPORTANT**:
116
+
- Treat each test case as an independently designed subtask with clear inputs/outputs.
117
+
- Execute independent subtasks in parallel when supported by the runtime.
118
+
- Pass only minimal required context to each subtask (target test case, relevant diff, and evaluation criteria) to keep context isolated.
119
+
- Normalize and aggregate subtask outputs in the final report.
116
120
117
121
#### Subagent Prompt
118
122
@@ -132,25 +136,29 @@ Read the prompt template from **`references/subagent-prompt.md`** and use it for
132
136
-`{test-case-diff}` - diff content for the specific test case
133
137
- Evaluates against 7 criteria (standard 6 + change appropriateness)
// In a single response, make Task tool calls for each modified test case:
152
-
Task(subagent_type="general-purpose", prompt="Review the golden test case 'interface-basic' modified in PR #123. Context: Changes in this PR... [diff-summary]. Test case diff: [test-case-diff]...")
153
-
```
152
+
Create one isolated subtask per modified test case (`1 test case = 1 isolated subtask`).
153
+
154
+
For each subtask, provide only:
155
+
- target test case name
156
+
- PR number or branch name
157
+
- diff summary limited to files relevant to that test case
158
+
- per-test-case diff
159
+
- required evaluation criteria
160
+
161
+
Do not share full PR/branch-wide cross-case context with every subtask. Cross-case reasoning (e.g., coverage gaps across categories) must be performed only in Step 3 aggregation.
154
162
155
163
### Step 3: Aggregate Results and Generate Report
156
164
@@ -183,6 +191,8 @@ After all subagents complete:
183
191
184
192
4.**Provide detailed reports** for test cases with issues (expand from JSON)
185
193
194
+
Note: Perform repository-wide or cross-case synthesis only at aggregation time, not inside isolated subtasks.
195
+
186
196
5.**(PR/Branch Review mode only) Include Test Coverage Analysis**:
Copy file name to clipboardExpand all lines: .claude/skills/test-pruning-advisor/SKILL.md
+11-4Lines changed: 11 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -13,7 +13,7 @@ Analyze test files to identify redundant, low-value, or obsolete tests that may
13
13
14
14
## Deletion Consideration Criteria
15
15
16
-
Based on project testing guidelines (CLAUDE.md):
16
+
Based on project testing guidelines (AGENTS.md / CLAUDE.md):
17
17
18
18
1.**Golden file test replaceable unit tests**: Function-level unit tests for code analysis, schema generation, or code generation logic that could be covered by testdata golden file tests
19
19
2.**Trivial tests**: Tests for simple getters/setters, direct value passthrough, or obvious behavior that provides minimal value
@@ -24,7 +24,7 @@ Based on project testing guidelines (CLAUDE.md):
24
24
25
25
### Step 1: Select Review Target
26
26
27
-
When the review target is not explicitly specified, use AskUserQuestion to present the following options:
27
+
When the review target is not explicitly specified, ask the user (using the runtime's available interaction mechanism) to choose from the following options:
28
28
29
29
1.**Branch/PR changes**: Review only test files added/modified in current branch vs main
30
30
2.**All tests**: Review all `*.test.ts` files in the codebase
@@ -60,9 +60,14 @@ Group test files by their purpose:
60
60
61
61
### Step 3: Launch Subagents for Analysis
62
62
63
-
For each test file (or batch of related files), launch a subagentusing the Task tool with `subagent_type: "general-purpose"`.
63
+
For each test file (or cohesive batch of closely related files), create a separate subtask with an isolated context (subagent/subtask model) using the runtime's available execution mechanism (`1 review unit = 1 isolated subtask`).
64
64
65
-
**IMPORTANT**: Launch all subagents in a single response with multiple Task tool calls to enable parallel execution.
65
+
**IMPORTANT**:
66
+
- Design each review unit as an independent subtask with explicit input/output.
67
+
- Execute independent subtasks in parallel when supported by the runtime.
68
+
- Pass only minimal required context to each subtask (target file(s), related source files, and evaluation criteria) to keep context isolated.
69
+
- Do not provide full repository-wide context to every subtask; keep context scoped to the review unit.
70
+
- Normalize and aggregate subtask outputs in the final report.
66
71
67
72
#### Subagent Prompt
68
73
@@ -91,6 +96,8 @@ After all subagents complete:
91
96
92
97
3.**List deletion candidates by reason**:
93
98
99
+
Note: Cross-file synthesis (e.g., overlap and priority decisions across multiple categories) must be done during aggregation, not inside isolated subtasks.
- Human review required each phase; use `-y` only for intentional fast-track
70
-
- Keep steering current and verify alignment with `/kiro:spec-status`
71
-
- Follow user instructions precisely; act autonomously within that scope; ask questions only when essential information is missing or instructions are critically ambiguous
72
-
73
-
**Language**: Think in English, generate responses in Japanese. All Markdown content written to project files (e.g., requirements.md, design.md, tasks.md) MUST be written in the target language configured for the specification (see spec.json.language).
0 commit comments