story-done▌
Donchitos/Claude-Code-Game-Studios · updated Apr 16, 2026
### Story Done
- ›description: "End-of-story completion review. Reads the story file, verifies each acceptance criterion against the implementation, checks for GDD/ADR deviations, prompts code review, updates story sta
- ›argument-hint: "[story-file-path] [--review full|lean|solo]"
- ›allowed-tools: Read, Glob, Grep, Bash, Edit, AskUserQuestion, Task
Story Done
This skill closes the loop between design and implementation. Run it at the end of implementing any story. It ensures every acceptance criterion is verified before the story is marked done, GDD and ADR deviations are explicitly documented rather than silently introduced, code review is prompted rather than forgotten, and the story file reflects actual completion status.
Output: Updated story file (Status: Complete) + surfaced next story.
Phase 1: Find the Story
Resolve the review mode (once, store for all gate spawns this run):
- If
--review [full|lean|solo]was passed → use that - Else read
production/review-mode.txt→ use that value - Else → default to
lean
See .claude/docs/director-gates.md for the full check pattern.
If a file path is provided (e.g., /story-done production/epics/core/story-damage-calculator.md):
read that file directly.
If no argument is provided:
- Check
production/session-state/active.mdfor the currently active story. - If not found there, read the most recent file in
production/sprints/and look for stories marked IN PROGRESS. - If multiple in-progress stories are found, use
AskUserQuestion:- "Which story are we completing?"
- Options: list the in-progress story file names.
- If no story can be found, ask the user to provide the path.
Phase 2: Read the Story
Read the full story file. Extract and hold in context:
- Story name and ID
- GDD Requirement TR-ID(s) referenced (e.g.,
TR-combat-001) - Manifest Version embedded in the story header (e.g.,
2026-03-10) - ADR reference(s) referenced
- Acceptance Criteria — the complete list (every checkbox item)
- Implementation files — files listed under "files to create/modify"
- Story Type — the
Type:field from the story header (Logic / Integration / Visual/Feel / UI / Config/Data) - Engine notes — any engine-specific constraints noted
- Definition of Done — if present, the story-level DoD
- Estimated vs actual scope — if an estimate was noted
Also read:
docs/architecture/tr-registry.yaml— look up each TR-ID in the story. Read the currentrequirementtext from the registry entry. This is the source of truth for what the GDD required — do not use any requirement text that may be quoted inline in the story (it may be stale).- The referenced GDD section — just the acceptance criteria and key rules, not the full document. Use this to cross-check the registry text is still accurate.
- The referenced ADR(s) — just the Decision and Consequences sections
docs/architecture/control-manifest.mdheader — extract the currentManifest Version:date (used in Phase 4 staleness check)
Phase 3: Verify Acceptance Criteria
For each acceptance criterion in the story, attempt verification using one of three methods:
Automatic verification (run without asking)
- File existence check:
Globfor files the story said would be created. - Test pass check: if a test file path is mentioned, run it via
Bash. - No hardcoded values check:
Grepfor numeric literals in gameplay code paths that should be in config files. - No hardcoded strings check:
Grepfor player-facing strings insrc/that should be in localization files. - Dependency check: if a criterion says "depends on X", check that X exists.
Manual verification with confirmation (use AskUserQuestion)
- Criteria about subjective qualities ("feels responsive", "animations play correctly")
- Criteria about gameplay behaviour ("player takes damage when...", "enemy responds to...")
- Performance criteria ("completes within Xms") — ask if profiled or accept as assumed
Batch up to 4 manual verification questions into a single AskUserQuestion call:
question: "Does [criterion]?"
options: "Yes — passes", "No — fails", "Not tested yet"
Unverifiable (flag without blocking)
- Criteria that require a full game build to test (end-to-end gameplay scenarios)
- Mark as:
DEFERRED — requires playtest session
Test-Criterion Traceability
After completing the pass/fail/deferred check above, map each acceptance criterion to the test that covers it:
For each acceptance criterion in the story:
-
Ask: is there a test — unit, integration, or confirmed manual playtest — that directly verifies this criterion?
- Unit test: check
tests/unit/for a test file or function name that matches the criterion's subject (useGlobandGrep) - Integration test: check
tests/integration/similarly - Manual confirmation: if the criterion was verified via
AskUserQuestionabove with a "Yes — passes" answer, count that as a manual test
- Unit test: check
-
Produce a traceability table:
| Criterion | Test | Status |
|-----------|------|--------|
| AC-1: [criterion text] | tests/unit/test_foo.gd::test_bar | COVERED |
| AC-2: [criterion text] | Manual playtest confirmation | COVERED |
| AC-3: [criterion text] | — | UNTESTED |
-
Apply these escalation rules:
- If >50% of criteria are UNTESTED: escalate to BLOCKING — test coverage is insufficient to confirm the story is actually done. The verdict in Phase 6 cannot be COMPLETE until coverage improves.
- If some (≤50%) criteria are UNTESTED: remain ADVISORY — does not block completion, but must appear in Completion Notes.
- If all criteria are COVERED: no action needed beyond including the table in the report.
-
For any ADVISORY untested criteria, add to the Completion Notes in Phase 7:
"Untested criteria: [AC-N list]. Recommend adding tests in a follow-up story."
Test Evidence Requirement
Based on the Story Type extracted in Phase 2, check for required evidence:
| Story Type | Required Evidence | Gate Level |
|---|---|---|
| Logic | Automated unit test in tests/unit/[system]/ — must exist and pass | BLOCKING |
| Integration | Integration test in tests/integration/[system]/ OR playtest doc | BLOCKING |
| Visual/Feel | Screenshot + sign-off in production/qa/evidence/ | ADVISORY |
| UI | Manual walkthrough doc OR interaction test in production/qa/evidence/ | ADVISORY |
| Config/Data | Smoke check pass report in production/qa/smoke-*.md | ADVISORY |
For Logic stories: first read the story's Test Evidence section to extract the
exact required file path. Use Glob to check that exact path. If the exact path is not
found, also search tests/unit/[system]/ broadly (the file may have been placed at a
slightly different location). If no test file is found at either location:
- Flag as BLOCKING: "Logic story has no unit test file. Story requires it at
[exact-path-from-Test-Evidence-section]. Create and run the test before marking this story Complete."
For Integration stories: read the story's Test Evidence section for the exact
required path. Use Glob to check that exact path first, then search
tests/integration/[system]/ broadly, then check production/session-logs/ for a
playtest record referencing this story.
If none found: flag as BLOCKING (same rule as Logic).
For Visual/Feel and UI stories: glob production/qa/evidence/ for a file
referencing this story. If none: flag as ADVISORY —
"No manual test evidence found. Create production/qa/evidence/[story-slug]-evidence.md
using the test-evidence template and obtain sign-off before final closure."
For Config/Data stories: check for any production/qa/smoke-*.md file.
If none: flag as ADVISORY — "No smoke check report found. Run /smoke-check."
If no Story Type is set: flag as ADVISORY —
"Story Type not declared. Add Type: [Logic|Integration|Visual/Feel|UI|Config/Data]
to the story header to enable test evidence gate enforcement in future stories."
Any BLOCKING test evidence gap prevents the COMPLETE verdict in Phase 6.
Phase 4: Check for Deviations
Compare the implementation against the design documents.
Run these checks automatically:
-
GDD rules check: Using the current requirement text from
tr-registry.yaml(looked up by the story's TR-ID), check that the implementation reflects what the GDD actually requires now — not what it required when the story was written.Grepthe implemented files for key function names, data structures, or class names mentioned in the current GDD section. -
Manifest version staleness check: Compare the
Manifest Version:date embedded in the story header against theManifest Version:date in the currentdocs/architecture/control-manifest.mdheader.- If they match → pass silently.
- If the story's version is older → flag as ADVISORY:
ADVISORY: Story was written against manifest v[story-date]; current manifest is v[current-date]. New rules may apply. Run /story-readiness to check. - If control-manifest.md does not exist → skip this check.
-
ADR constraints check: Read the referenced ADR's Decision section. Check for forbidden patterns from
docs/architecture/control-manifest.md(if it exists).Grepfor patterns explicitly forbidden in the ADR. -
Hardcoded values check:
Grepthe implemented files for numeric literals in gameplay logic that should be in data files. -
Scope check: Did the implementation touch files outside the story's stated scope? (files not listed in "files to create/modify")
For each deviation found, categorize:
- BLOCKING — implementation contradicts the GDD or ADR (must fix before marking complete)
- ADVISORY — implementation drifts slightly from spec but is functionally equivalent (document, user decides)
- OUT OF SCOPE — additional files were touched beyond the story's stated boundary (flag for awareness — may be valid or scope creep)
Phase 4b: QA Coverage Gate
Review mode check — apply before spawning QL-TEST-COVERAGE:
solo→ skip. Note: "QL-TEST-COVERAGE skipped — Solo mode." Proceed to Phase 5.lean→ skip (not a PHASE-GATE). Note: "QL-TEST-COVERAGE skipped — Lean mode." Proceed to Phase 5.full→ spawn as normal.
After completing the deviation checks in Phase 4, spawn qa-lead via Task using gate QL-TEST-COVERAGE (.claude/docs/director-gates.md).
Pass:
- The story file path and story type
- Test file paths found during Phase 3 (exact paths, or "none found")
- The story's
## QA Test Casessection (the pre-written test specs from story creation) - The story's
## Acceptance Criterialist
The qa-lead reviews whether the tests actually cover what was specified — not just whether files exist.
Apply the verdict:
- ADEQUATE → proceed to Phase 5
- GAPS → flag as ADVISORY: "QA lead identified coverage gaps: [list]. Story can complete but gaps should be addressed in a follow-up story."
- INADEQUATE → flag as BLOCKING: "QA lead: critical logic is untested. Verdict cannot be COMPLETE until coverage improves. Specific gaps: [list]."
Skip this phase for Config/Data stories (no code tests required).
Phase 5: Lead Programmer Code Review Gate
Review mode check — apply before spawning LP-CODE-REVIEW:
solo→ skip. Note: "LP-CODE-REVIEW skipped — Solo mode." Proceed to Phase 6 (completion report).lean→ skip (not a PHASE-GATE). Note: "LP-CODE-REVIEW skipped — Lean mode." Proceed to Phase 6 (completion report).full→ spawn as normal.
Spawn lead-programmer via Task using gate LP-CODE-REVIEW (.claude/docs/director-gates.md).
Pass: implementation file paths, story file path, relevant GDD section, governing ADR.
Present the verdict to the user. If CONCERNS, surface them via AskUserQuestion:
- Options:
Revise flagged issues/Accept and proceed/Discuss furtherIf REJECT, do not proceed to Phase 6 verdict until the issues are resolved.
If the story has no implementation files yet (verdict is being run before coding is done), skip this phase and note: "LP-CODE-REVIEW skipped — no implementation files found. Run after implementation is complete."
Phase 6: Present the Completion Report
Before updating any files, present the full report:
## Story Done: [Story Name]
**Story**: [file path]
**Date**: [today]
### Acceptance Criteria: [X/Y passing]
- [x] [Criterion 1] — auto-verified (test passes)
- [x] [Criterion 2] — confirmed
- [ ] [Criterion 3] — FAILS: [reason]
- [?] [Criterion 4] — DEFERRED: requires playtest
### Test-Criterion Traceability
| Criterion | Test | Status |
|-----------|------|--------|
| AC-1: [text] | [test file::test name] | COVERED |
| AC-2: [text] | Manual confirmation | COVERED |
| AC-3: [text] | — | UNTESTED |
### Test Evidence
**Story Type**: [Logic | Integration | Visual/Feel | UI | Config/Data | Not declared]
**Required evidence**: [unit test file | integration test or playtest | screenshot + sign-off | walkthrough doc | smoke check pass]
**Evidence found**: [YES — `[path]` | NO — BLOCKING | NO — ADVISORY]
### Deviations
[NONE] OR:
- BLOCKING: [description] — [GDD/ADR reference]
- ADVISORY: [description] — user accepted / flagged for tech debt
### Scope
[All changes within stated scope] OR:
- Extra files touched: [list] — [note whether valid or scope creep]
### Verdict: COMPLETE / COMPLETE WITH NOTES / BLOCKED
Verdict definitions:
- COMPLETE: all criteria pass, no blocking deviations
- COMPLETE WITH NOTES: all criteria pass, advisory deviations documented
- BLOCKED: failing criteria or blocking deviations must be resolved first
If the verdict is BLOCKED: do not proceed to Phase 7. List what must be fixed. Offer to help fix the blocking items.
Phase 7: Update Story Status
Ask before writing: "May I update the story file to mark it Complete and log the completion notes?"
If yes, edit the story file:
- Update the status field:
Status: Complete - Add a
## Completion Notessection at the bottom:
## Completion Notes
**Completed**: [date]
**Criteria**: [X/Y passing] ([any deferred items listed])
**Deviations**: [None] or [list of advisory deviations]
**Test Evidence**: [Logic: test file at path | Visual/Feel: evidence doc at path | None required (Config/Data)]
**Code Review**: [Pending / Complete / Skipped]
-
If advisory deviations exist, ask: "Should I log these as tech debt in
docs/tech-debt-register.md?" -
Update
production/sprint-status.yaml(if it exists):- Find the entry matching this story's file path or ID
- Set
status: doneandcompleted: [today's date] - Update the top-level
updatedfield - This is a silent update — no extra approval needed (already approved in step above)
Session State Update
After updating the story file, silently append to
production/session-state/active.md:
## Session Extract — /story-done [date]
- Verdict: [COMPLETE / COMPLETE WITH NOTES / BLOCKED]
- Story: [story file path] — [story title]
- Tech debt logged: [N items, or "None"]
- Next recommended: [next ready story title and path, or "None identified"]
If active.md does not exist, create it with this block as the initial content.
Confirm in conversation: "Session state updated."
Phase 8: Surface the Next Story
After completion, help the developer keep momentum:
- Read the current sprint plan from
production/sprints/. - Find stories that are:
- Status: READY or NOT STARTED
- Not blocked by other incomplete stories
- In the Must Have or Should Have tier
Present:
### Next Up
The following stories are ready to pick up:
1. [Story name] — [1-line description] — Est: [X hrs]
2. [Story name] — [1-line description] — Est: [X hrs]
Run `/story-readiness [path]` to confirm a story is implementation-ready
before starting.
If no more Must Have stories remain in this sprint (all are Complete or Blocked):
### Sprint Close-Out Sequence
All Must Have stories are complete. QA sign-off is required before advancing.
Run these in order:
1. `/smoke-check sprint` — verify the critical path still works end-to-end
2. `/team-qa sprint` — full QA cycle: test case execution, bug triage, sign-off report
3. `/gate-check` — advance to the next phase once QA approves
Do not run `/gate-check` until `/team-qa` returns APPROVED or APPROVED WITH CONDITIONS.
If there are Should Have stories still unstarted, surface them alongside the close-out sequence so the user can choose: close the sprint now, or pull in more work first.
If no more stories are ready but Must Have stories are still In Progress (not Complete): "No more stories ready to start — [N] Must Have stories still in progress. Continue implementing those before sprint close-out."
Collaborative Protocol
- Never mark a story complete without user approval — Phase 7 requires an explicit "yes" before any file is edited.
- Never auto-fix failing criteria — report them and ask what to do.
- Deviations are facts, not judgments — present them neutrally; the user decides if they are acceptable.
- BLOCKED verdict is advisory — the user can override and mark complete anyway; document the risk explicitly if they do.
- Use
AskUserQuestionfor the code review prompt and for batching manual criteria confirmations.
Recommended Next Steps
- Run
/story-readiness [next-story-path]to validate the next story before starting implementation - If all Must Have stories are complete: run
/smoke-check sprint→/team-qa sprint→/gate-check - If tech debt was logged: track it via
/tech-debtto keep the register current
Ratings
4.5★★★★★44 reviews- ★★★★★Yuki Bhatia· Dec 24, 2024
story-done has been reliable in day-to-day use. Documentation quality is above average for community skills.
- ★★★★★Yuki Bansal· Dec 20, 2024
I recommend story-done for anyone iterating fast on agent tooling; clear intent and a small, reviewable surface area.
- ★★★★★Layla Yang· Dec 16, 2024
Keeps context tight: story-done is the kind of skill you can hand to a new teammate without a long onboarding doc.
- ★★★★★Rahul Santra· Nov 15, 2024
We added story-done from the explainx registry; install was straightforward and the SKILL.md answered most questions upfront.
- ★★★★★Aarav Rao· Nov 15, 2024
We added story-done from the explainx registry; install was straightforward and the SKILL.md answered most questions upfront.
- ★★★★★Meera Huang· Nov 15, 2024
Useful defaults in story-done — fewer surprises than typical one-off scripts, and it plays nicely with `npx skills` flows.
- ★★★★★Aanya Shah· Nov 11, 2024
Solid pick for teams standardizing on skills: story-done is focused, and the summary matches what you get after install.
- ★★★★★Kiara Kapoor· Nov 7, 2024
Registry listing for story-done matched our evaluation — installs cleanly and behaves as described in the markdown.
- ★★★★★Daniel Thompson· Oct 26, 2024
story-done reduced setup friction for our internal harness; good balance of opinion and flexibility.
- ★★★★★Pratham Ware· Oct 6, 2024
story-done fits our agent workflows well — practical, well scoped, and easy to wire into existing repos.
showing 1-10 of 44