architecture-review▌
Donchitos/Claude-Code-Game-Studios · updated Apr 16, 2026
### Architecture Review
- ›name: architecture-review
- ›description: "Validates completeness and consistency of the project architecture against all GDDs. Builds a traceability matrix mapping every GDD technical requirement to ADRs, identifies coverage gap
- ›argument-hint: "[focus: full | coverage | consistency | engine | single-gdd path/to/gdd.md]"
Architecture Review
The architecture review validates that the complete body of architectural decisions covers all game design requirements, is internally consistent, and correctly targets the project's pinned engine version. It is the quality gate between Technical Setup and Pre-Production.
Argument modes:
- No argument /
full: Full review — all phases coverage: Traceability only — which GDD requirements have no ADRconsistency: Cross-ADR conflict detection onlyengine: Engine compatibility audit onlysingle-gdd [path]: Review architecture coverage for one specific GDDrtm: Requirements Traceability Matrix — extends the standard matrix to include story file paths and test file paths; outputsdocs/architecture/requirements-traceability.mdwith the full GDD requirement → ADR → Story → Test chain. Use in Production phase when stories and tests exist.
Phase 1: Load Everything
Phase 1a — L0: Summary Scan (fast, low tokens)
Before reading any full document, use Grep to extract ## Summary sections
from all GDDs and ADRs:
Grep pattern="## Summary" glob="design/gdd/*.md" output_mode="content" -A 4
Grep pattern="## Summary" glob="docs/architecture/adr-*.md" output_mode="content" -A 3
For single-gdd [path] mode: use the target GDD's summary to identify which
ADRs reference the same system (Grep ADRs for the system name), then full-read
only those ADRs. Skip full-reading unrelated GDDs entirely.
For engine mode: only full-read ADRs — GDDs are not needed for engine checks.
For coverage or full mode: proceed to full-read everything below.
Phase 1b — L1/L2: Full Document Load
Read all inputs appropriate to the mode:
Design Documents
- All in-scope GDDs in
design/gdd/— read every file completely design/gdd/systems-index.md— the authoritative list of systems
Architecture Documents
- All in-scope ADRs in
docs/architecture/— read every file completely docs/architecture/architecture.mdif it exists
Engine Reference
docs/engine-reference/[engine]/VERSION.mddocs/engine-reference/[engine]/breaking-changes.mddocs/engine-reference/[engine]/deprecated-apis.md- All files in
docs/engine-reference/[engine]/modules/
Project Standards
.claude/docs/technical-preferences.md
Report a count: "Loaded [N] GDDs, [M] ADRs, engine: [name + version]."
Also read docs/consistency-failures.md if it exists. Extract entries with
Domain matching the systems under review (Architecture, Engine, or any GDD domain
being covered). Surface recurring patterns as a "Known conflict-prone areas" note
at the top of the Phase 4 conflict detection output.
Phase 2: Extract Technical Requirements from Every GDD
Pre-load the TR Registry
Before extracting any requirements, read docs/architecture/tr-registry.yaml
if it exists. Index existing entries by id and by normalized requirement
text (lowercase, trimmed). This prevents ID renumbering across review runs.
For each requirement you extract, the matching rule is:
- Exact/near match to an existing registry entry for the same system →
reuse that entry's TR-ID unchanged. Update the
requirementtext in the registry only if the GDD wording changed (same intent, clearer phrasing) — add arevised: [date]field. - No match → assign a new ID: next available
TR-[system]-NNNfor that system, starting from the highest existing sequence + 1. - Ambiguous (partial match, intent unclear) → ask the user:
"Does '[new requirement text]' refer to the same requirement as
TR-[system]-NNN: [existing text]', or is it a new requirement?" User answers: "Same requirement" (reuse ID) or "New requirement" (new ID).
For any requirement with status: deprecated in the registry — skip it.
It was removed from the GDD intentionally.
For each GDD, read it and extract all technical requirements — things the architecture must provide for the system to work. A technical requirement is any statement that implies a specific architectural decision.
Categories to extract:
| Category | Example |
|---|---|
| Data structures | "Each entity has health, max health, status effects" → needs a component/data schema |
| Performance constraints | "Collision detection must run at 60fps with 200 entities" → physics budget ADR |
| Engine capability | "Inverse kinematics for character animation" → IK system ADR |
| Cross-system communication | "Damage system notifies UI and audio simultaneously" → event/signal architecture ADR |
| State persistence | "Player progress persists between sessions" → save system ADR |
| Threading/timing | "AI decisions happen off the main thread" → concurrency ADR |
| Platform requirements | "Supports keyboard, gamepad, touch" → input system ADR |
For each GDD, produce a structured list:
GDD: [filename]
System: [system name]
Technical Requirements:
TR-[GDD]-001: [requirement text] → Domain: [Physics/Rendering/etc]
TR-[GDD]-002: [requirement text] → Domain: [...]
This becomes the requirements baseline — the complete set of what the architecture must cover.
Phase 3: Build the Traceability Matrix
For each technical requirement extracted in Phase 2, search the ADRs:
- Read every ADR's "GDD Requirements Addressed" section
- Check if it explicitly references the requirement or its GDD
- Check if the ADR's decision text implicitly covers the requirement
- Mark coverage status:
| Status | Meaning |
|---|---|
| ✅ Covered | An ADR explicitly addresses this requirement |
| ⚠️ Partial | An ADR partially covers this, or coverage is ambiguous |
| ❌ Gap | No ADR addresses this requirement |
Build the full matrix:
## Traceability Matrix
| Requirement ID | GDD | System | Requirement | ADR Coverage | Status |
|---------------|-----|--------|-------------|--------------|--------|
| TR-combat-001 | combat.md | Combat | Hitbox detection < 1 frame | ADR-0003 | ✅ |
| TR-combat-002 | combat.md | Combat | Combo window timing | — | ❌ GAP |
| TR-inventory-001 | inventory.md | Inventory | Persistent item storage | ADR-0005 | ✅ |
Count the totals: X covered, Y partial, Z gaps.
Phase 3b: Story and Test Linkage (RTM mode only)
Skip this phase unless the argument is rtm or full with stories present.
This phase extends the Phase 3 matrix to include the story that implements each requirement and the test that verifies it — producing the full Requirements Traceability Matrix (RTM).
Step 3b-1 — Load stories
Glob production/epics/**/*.md (excluding EPIC.md index files). For each
story file:
- Extract
TR-IDfrom the story's Context section - Extract story file path, title, Status
- Extract
## Test Evidencesection — the stated test file path
Step 3b-2 — Load test files
Glob tests/unit/**/*_test.* and tests/integration/**/*_test.*.
Build an index: system → [test file paths].
For each test file path from Step 3b-1, confirm via Glob whether the file actually exists. Note MISSING if the stated path does not exist.
Step 3b-3 — Build the extended RTM
For each TR-ID in the Phase 3 matrix, add:
- Story: the story file path(s) that reference this TR-ID (may be multiple)
- Test File: the test file path stated in the story's Test Evidence section
- Test Status: COVERED (test file exists) / MISSING (path stated but not found) / NONE (no test path stated, story type may be Visual/Feel/UI) / NO STORY (requirement has no story yet — pre-production gap)
Extended matrix format:
## Requirements Traceability Matrix (RTM)
| TR-ID | GDD | Requirement | ADR | Story | Test File | Test Status |
|-------|-----|-------------|-----|-------|-----------|-------------|
| TR-combat-001 | combat.md | Hitbox < 1 frame | ADR-0003 | story-001-hitbox.md | tests/unit/combat/hitbox_test.gd | COVERED |
| TR-combat-002 | combat.md | Combo window | — | story-002-combo.md | — | NONE (Visual/Feel) |
| TR-inventory-001 | inventory.md | Persistent storage | ADR-0005 | — | — | NO STORY |
RTM coverage summary:
- COVERED: [N] — requirements with ADR + story + passing test
- MISSING test: [N] — story exists but test file not found
- NO STORY: [N] — requirements with ADR but no story yet
- NO ADR: [N] — requirements without architectural coverage (from Phase 3 gaps)
- Full chain complete (COVERED): [N/total] ([%])
Phase 4: Cross-ADR Conflict Detection
Compare every ADR against every other ADR to detect contradictions. A conflict exists when:
- Data ownership conflict: Two ADRs claim exclusive ownership of the same data
- Integration contract conflict: ADR-A assumes System X has interface Y, but ADR-B defines System X with a different interface
- Performance budget conflict: ADR-A allocates N ms to physics, ADR-B allocates N ms to AI, together they exceed the total frame budget
- Dependency cycle: ADR-A says System X initialises before Y; ADR-B says Y initialises before X
- Architecture pattern conflict: ADR-A uses event-driven communication for a subsystem; ADR-B uses direct function calls to the same subsystem
- State management conflict: Two ADRs define authority over the same game state (e.g. both Combat ADR and Character ADR claim to own the health value)
For each conflict found:
## Conflict: [ADR-NNNN] vs [ADR-MMMM]
Type: [Data ownership / Integration / Performance / Dependency / Pattern / State]
ADR-NNNN claims: [...]
ADR-MMMM claims: [...]
Impact: [What breaks if both are implemented as written]
Resolution options:
1. [Option A]
2. [Option B]
ADR Dependency Ordering
After conflict detection, analyse the dependency graph across all ADRs:
- Collect all
Depends Onfields from every ADR's "ADR Dependencies" section - Topological sort: Determine the correct implementation order — ADRs with no dependencies come first (Foundation), ADRs that depend on those come next, etc.
- Flag unresolved dependencies: If ADR-A's "Depends On" field references an ADR
that is still
Proposedor does not exist, flag it:⚠️ ADR-0005 depends on ADR-0002 — but ADR-0002 is still Proposed. ADR-0005 cannot be safely implemented until ADR-0002 is Accepted. - Cycle detection: If ADR-A depends on ADR-B and ADR-B depends on ADR-A (directly
or transitively), flag it as a
DEPENDENCY CYCLE:🔴 DEPENDENCY CYCLE: ADR-0003 → ADR-0006 → ADR-0003 This cycle must be broken before either can be implemented. - Output recommended implementation order:
### Recommended ADR Implementation Order (topologically sorted) Foundation (no dependencies): 1. ADR-0001: [title] 2. ADR-0003: [title] Depends on Foundation: 3. ADR-0002: [title] (requires ADR-0001) 4. ADR-0005: [title] (requires ADR-0003) Feature layer: 5. ADR-0004: [title] (requires ADR-0002, ADR-0005)
Phase 5: Engine Compatibility Cross-Check
Across all ADRs, check for engine consistency:
Version Consistency
- Do all ADRs that mention an engine version agree on the same version?
- If any ADR was written for an older engine version, flag it as potentially stale
Post-Cutoff API Consistency
- Collect all "Post-Cutoff APIs Used" fields from all ADRs
- For each, verify against the relevant module reference doc
- Check that no two ADRs make contradictory assumptions about the same post-cutoff API
Deprecated API Check
- Grep all ADRs for API names listed in
deprecated-apis.md - Flag any ADR referencing a deprecated API
Missing Engine Compatibility Sections
- List all ADRs that are missing the Engine Compatibility section entirely
- These are blind spots — their engine assumptions are unknown
Output format:
### Engine Audit Results
Engine: [name + version]
ADRs with Engine Compatibility section: X / Y total
Deprecated API References:
- ADR-0002: uses [deprecated API] — deprecated since [version]
Stale Version References:
- ADR-0001: written for [older version] — current project version is [version]
Post-Cutoff API Conflicts:
- ADR-0004 and ADR-0007 both use [API] with incompatible assumptions
Engine Specialist Consultation
After completing the engine audit above, spawn the primary engine specialist via Task for a domain-expert second opinion:
- Read
.claude/docs/technical-preferences.mdEngine Specialistssection to get the primary specialist - If no engine is configured, skip this consultation
- Spawn
subagent_type: [primary specialist]with: all ADRs that contain engine-specific decisions orPost-Cutoff APIs Usedfields, the engine reference docs, and the Phase 5 audit findings. Ask them to:- Confirm or challenge each audit finding — specialists may know of engine nuances not captured in the reference docs
- Identify engine-specific anti-patterns in the ADRs that the audit may have missed (e.g., using the wrong Godot node type, Unity component coupling, Unreal subsystem misuse)
- Flag ADRs that make assumptions about engine behaviour that differ from the actual pinned version
Incorporate additional findings under ### Engine Specialist Findings in the Phase 5 output. These feed into the final verdict — specialist-identified issues carry the same weight as audit-identified issues.
Phase 5b: Design Revision Flags (Architecture → GDD Feedback)
For each HIGH RISK engine finding from Phase 5, check whether any GDD makes an assumption that the verified engine reality contradicts.
Specific cases to check:
-
Post-cutoff API behaviour differs from training-data assumptions: If an ADR records a verified API behaviour that differs from the default LLM assumption, check all GDDs that reference the related system. Look for design rules written around the old (assumed) behaviour.
-
Known engine limitations in ADRs: If an ADR records a known engine limitation (e.g. "Jolt ignores HingeJoint3D damp", "D3D12 is now the default backend"), check GDDs that design mechanics around the affected feature.
-
Deprecated API conflicts: If Phase 5 flagged a deprecated API used in an ADR, check whether any GDD contains mechanics that assume the deprecated API's behaviour.
For each conflict found, record it in the GDD Revision Flags table:
### GDD Revision Flags (Architecture → Design Feedback)
These GDD assumptions conflict with verified engine behaviour or accepted ADRs.
The GDD should be revised before its system enters implementation.
| GDD | Assumption | Reality (from ADR/engine-reference) | Action |
|-----|-----------|--------------------------------------|--------|
| combat.md | "Use HingeJoint3D damp for weapon recoil" | Jolt ignores damp — ADR-0003 | Revise GDD |
If no revision flags are found, write: "No GDD revision flags — all GDD assumptions are consistent with verified engine behaviour."
Ask: "Should I flag these GDDs for revision in the systems index?"
- If yes: update the relevant systems' Status field to "Needs Revision" and add a short inline note in the adjacent Notes/Description column explaining the conflict. Ask for approval before writing. (Do NOT use parentheticals like "Needs Revision (Architecture Feedback)" — other skills match the exact string "Needs Revision" and parentheticals break that match.)
Phase 6: Architecture Document Coverage
If docs/architecture/architecture.md exists, validate it against GDDs:
- Does every system from
systems-index.mdappear in the architecture layers? - Does the data flow section cover all cross-system communication defined in GDDs?
- Do the API boundaries support all integration requirements from GDDs?
- Are there systems in the architecture doc that have no corresponding GDD (orphaned architecture)?
Phase 7: Output the Review Report
## Architecture Review Report
Date: [date]
Engine: [name + version]
GDDs Reviewed: [N]
ADRs Reviewed: [M]
---
### Traceability Summary
Total requirements: [N]
✅ Covered: [X]
⚠️ Partial: [Y]
❌ Gaps: [Z]
### Coverage Gaps (no ADR exists)
For each gap:
❌ TR-[id]: [GDD] → [system] → [requirement]
Suggested ADR: "/architecture-decision [suggested title]"
Domain: [Physics/Rendering/etc]
Engine Risk: [LOW/MEDIUM/HIGH]
### Cross-ADR Conflicts
[List all conflicts from Phase 4]
### ADR Dependency Order
[Topologically sorted implementation order from Phase 4 — dependency ordering section]
[Unresolved dependencies and cycles if any]
### GDD Revision Flags
[GDD assumptions that conflict with verified engine behaviour — from Phase 5b]
[Or: "None — all GDD assumptions consistent with verified engine behaviour"]
### Engine Compatibility Issues
[List all engine issues from Phase 5]
### Architecture Document Coverage
[List missing systems and orphaned architecture from Phase 6]
---
### Verdict: [PASS / CONCERNS / FAIL]
PASS: All requirements covered, no conflicts, engine consistent
CONCERNS: Some gaps or partial coverage, but no blocking conflicts
FAIL: Critical gaps (Foundation/Core layer requirements uncovered),
or blocking cross-ADR conflicts detected
### Blocking Issues (must resolve before PASS)
[List items that must be resolved — FAIL verdict only]
### Required ADRs
[Prioritised list of ADRs to create, most foundational first]
Phase 8: Write and Update Traceability Index
Use AskUserQuestion for the write approval:
- "Review complete. What would you like to write?"
- [A] Write all three files (review report + traceability index + TR registry)
- [B] Write review report only —
docs/architecture/architecture-review-[date].md - [C] Don't write anything yet — I need to review the findings first
RTM Output (rtm mode only)
For rtm mode, additionally ask: "May I write the full Requirements Traceability
Matrix to docs/architecture/requirements-traceability.md?"
RTM file format:
# Requirements Traceability Matrix (RTM)
> Last Updated: [date]
> Mode: /architecture-review rtm
> Coverage: [N]% full chain complete (GDD → ADR → Story → Test)
## How to read this matrix
| Column | Meaning |
|--------|---------|
| TR-ID | Stable requirement ID from tr-registry.yaml |
| GDD | Source design document |
| ADR | Architectural decision governing implementation |
| Story | Story file that implements this requirement |
| Test File | Automated test file path |
| Test Status | COVERED / MISSING / NONE / NO STORY |
## Full Traceability Matrix
| TR-ID | GDD | Requirement | ADR | Story | Test File | Status |
|-------|-----|-------------|-----|-------|-----------|--------|
[Full matrix rows from Phase 3b]
## Coverage Summary
| Status | Count | % |
|--------|-------|---|
| COVERED — full chain complete | [N] | [%] |
| MISSING test — story exists, no test | [N] | [%] |
| NO STORY — ADR exists, not yet implemented | [N] | [%] |
| NO ADR — architectural gap | [N] | [%] |
| **Total requirements** | **[N]** | **100%** |
## Uncovered Requirements (Priority Fix List)
Requirements where the full chain is broken, prioritised by layer:
### Foundation layer gaps
[list with suggested action per gap]
### Core layer gaps
[list]
### Feature / Presentation layer gaps
[list — lower priority]
## History
| Date | Full Chain % | Notes |
|------|-------------|-------|
| [date] | [%] | Initial RTM |
TR Registry Update
Also ask: "May I update docs/architecture/tr-registry.yaml with new requirement
IDs from this review?"
If yes:
- Append any new TR-IDs that weren't in the registry before this review
- Update
requirementtext andreviseddate for any entries whose GDD wording changed (ID stays the same) - Mark
status: deprecatedfor any registry entries whose GDD requirement no longer exists (confirm with user before marking deprecated) - Never renumber or delete existing entries
- Update the
last_updatedandversionfields at the top
This ensures all future story files can reference stable TR-IDs that persist across every subsequent architecture review.
Reflexion Log Update
After writing the review report, append any 🔴 CONFLICT entries found in Phase 4
to docs/consistency-failures.md (if the file exists):
### [YYYY-MM-DD] — /architecture-review — 🔴 CONFLICT
**Domain**: Architecture / [specific domain e.g. State Ownership, Performance]
**Documents involved**: [ADR-NNNN] vs [ADR-MMMM]
**What happened**: [specific conflict — what each ADR claims]
**Resolution**: [how it was or should be resolved]
**Pattern**: [generalised lesson for future ADR authors in this domain]
Only append CONFLICT entries — do not log GAP entries (missing ADRs are expected before the architecture is complete). Do not create the file if missing — only append when it already exists.
Session State Update
After writing all approved files, silently append to
production/session-state/active.md:
## Session Extract — /architecture-review [date]
- Verdict: [PASS / CONCERNS / FAIL]
- Requirements: [N] total — [X] covered, [Y] partial, [Z] gaps
- New TR-IDs registered: [N, or "None"]
- GDD revision flags: [comma-separated GDD names, or "None"]
- Top ADR gaps: [top 3 gap titles from the report, or "None"]
- Report: docs/architecture/architecture-review-[date].md
If active.md does not exist, create it with this block as the initial content.
Confirm in conversation: "Session state updated."
The traceability index format:
# Architecture Traceability Index
Last Updated: [date]
Engine: [name + version]
## Coverage Summary
- Total requirements: [N]
- Covered: [X] ([%])
- Partial: [Y]
- Gaps: [Z]
## Full Matrix
[Complete traceability matrix from Phase 3]
## Known Gaps
[All ❌ items with suggested ADRs]
## Superseded Requirements
[Requirements whose GDD was changed after the ADR was written]
Phase 9: Handoff
After completing the review and writing approved files, present:
- Immediate actions: List the top 3 ADRs to create (highest-impact gaps first, Foundation layer before Feature layer)
- Gate guidance: "When all blocking issues are resolved, run
/gate-check pre-productionto advance" - Rerun trigger: "Re-run
/architecture-reviewafter each new ADR is written to verify coverage improves"
Then close with AskUserQuestion:
- "Architecture review complete. What would you like to do next?"
- [A] Write a missing ADR — open a fresh session and run
/architecture-decision [system] - [B] Run
/gate-check pre-production— if all blocking gaps are resolved - [C] Stop here for this session
- [A] Write a missing ADR — open a fresh session and run
Error Recovery Protocol
If any spawned agent returns BLOCKED, errors, or fails to complete:
- Surface immediately: Report "[AgentName]: BLOCKED — [reason]" before continuing
- Assess dependencies: If the blocked agent's output is required by a later phase, do not proceed past that phase without user input
- Offer options via AskUserQuestion with three choices:
- Skip this agent and note the gap in the final report
- Retry with narrower scope (fewer GDDs, single-system focus)
- Stop here and resolve the blocker first
- Always produce a partial report — output whatever was completed so work is not lost
Collaborative Protocol
- Read silently — do not narrate every file read
- Show the matrix — present the full traceability matrix before asking for anything; let the user see the state
- Don't guess — if a requirement is ambiguous, ask: "Is [X] a technical requirement or a design preference?"
- Ask before writing — always confirm before writing the report file
- Non-blocking — the verdict is advisory; the user decides whether to continue despite CONCERNS or even FAIL findings
Ratings
4.5★★★★★64 reviews- ★★★★★Pratham Ware· Dec 20, 2024
Registry listing for architecture-review matched our evaluation — installs cleanly and behaves as described in the markdown.
- ★★★★★William Li· Dec 8, 2024
Keeps context tight: architecture-review is the kind of skill you can hand to a new teammate without a long onboarding doc.
- ★★★★★Sakura Martin· Dec 4, 2024
I recommend architecture-review for anyone iterating fast on agent tooling; clear intent and a small, reviewable surface area.
- ★★★★★Ama Liu· Nov 27, 2024
I recommend architecture-review for anyone iterating fast on agent tooling; clear intent and a small, reviewable surface area.
- ★★★★★Hiroshi Li· Nov 23, 2024
Keeps context tight: architecture-review is the kind of skill you can hand to a new teammate without a long onboarding doc.
- ★★★★★Sakshi Patil· Nov 11, 2024
architecture-review reduced setup friction for our internal harness; good balance of opinion and flexibility.
- ★★★★★Kabir Gonzalez· Oct 18, 2024
architecture-review reduced setup friction for our internal harness; good balance of opinion and flexibility.
- ★★★★★Advait Shah· Oct 14, 2024
Registry listing for architecture-review matched our evaluation — installs cleanly and behaves as described in the markdown.
- ★★★★★Chaitanya Patil· Oct 2, 2024
I recommend architecture-review for anyone iterating fast on agent tooling; clear intent and a small, reviewable surface area.
- ★★★★★Kwame Ramirez· Sep 25, 2024
architecture-review has been reliable in day-to-day use. Documentation quality is above average for community skills.
showing 1-10 of 64