EasyPlatform cook
[IMPORTANT]** Use `TaskCreate` to break ALL work into small tasks BEFORE starting — including tasks for each file read. This prevents context loss from long files. For simple tasks, AI MUST ATTENTION ask user whether to skip.
git clone https://github.com/duc01226/EasyPlatform
T=$(mktemp -d) && git clone --depth=1 https://github.com/duc01226/EasyPlatform "$T" && mkdir -p ~/.claude/skills && cp -r "$T/.claude/skills/cook" ~/.claude/skills/duc01226-easyplatform-cook && rm -rf "$T"
.claude/skills/cook/SKILL.md<!-- SYNC:critical-thinking-mindset -->[IMPORTANT] Use
to break ALL work into small tasks BEFORE starting — including tasks for each file read. This prevents context loss from long files. For simple tasks, AI MUST ATTENTION ask user whether to skip.TaskCreate
<!-- /SYNC:critical-thinking-mindset --> <!-- SYNC:ai-mistake-prevention -->Critical Thinking Mindset — Apply critical thinking, sequential thinking. Every claim needs traced proof, confidence >80% to act. Anti-hallucination: Never present guess as fact — cite sources for every claim, admit uncertainty freely, self-check output for errors, cross-reference independently, stay skeptical of own confidence — certainty without evidence root of all hallucination.
<!-- /SYNC:ai-mistake-prevention --> <!-- SYNC:understand-code-first -->AI Mistake Prevention — Failure modes to avoid on every task:
- Check downstream references before deleting. Deleting components causes documentation and code staleness cascades. Map all referencing files before removal.
- Verify AI-generated content against actual code. AI hallucinates APIs, class names, and method signatures. Always grep to confirm existence before documenting or referencing.
- Trace full dependency chain after edits. Changing a definition misses downstream variables and consumers derived from it. Always trace the full chain.
- Trace ALL code paths when verifying correctness. Confirming code exists is not confirming it executes. Always trace early exits, error branches, and conditional skips — not just happy path.
- When debugging, ask "whose responsibility?" before fixing. Trace whether bug is in caller (wrong data) or callee (wrong handling). Fix at responsible layer — never patch symptom site.
- Assume existing values are intentional — ask WHY before changing. Before changing any constant, limit, flag, or pattern: read comments, check git blame, examine surrounding code.
- Verify ALL affected outputs, not just the first. Changes touching multiple stacks require verifying EVERY output. One green check is not all green checks.
- Holistic-first debugging — resist nearest-attention trap. When investigating any failure, list EVERY precondition first (config, env vars, DB names, endpoints, DI registrations, data preconditions), then verify each against evidence before forming any code-layer hypothesis.
- Surgical changes — apply the diff test. Bug fix: every changed line must trace directly to the bug. Don't restyle or improve adjacent code. Enhancement task: implement improvements AND announce them explicitly.
- Surface ambiguity before coding — don't pick silently. If request has multiple interpretations, present each with effort estimate and ask. Never assume all-records, file-based, or more complex path.
<!-- /SYNC:understand-code-first --> <!-- SYNC:double-round-trip-review -->Understand Code First — HARD-GATE: Do NOT write, plan, or fix until you READ existing code.
- Search 3+ similar patterns (
/grep) — citeglobevidencefile:line- Read existing files in target area — understand structure, base classes, conventions
- Run
whenpython .claude/scripts/code_graph trace <file> --direction both --jsonexists.code-graph/graph.db- Map dependencies via
orconnections— know what depends on your targetcallers_of- Write investigation to
for non-trivial tasks (3+ files).ai/workspace/analysis/- Re-read analysis file before implementing — never work from memory alone
- NEVER invent new patterns when existing ones work — match exactly or document deviation
BLOCKED until:
Read target files- [ ]Grep 3+ patterns- [ ]Graph trace (if graph.db exists)- [ ]Assumptions verified with evidence- [ ]
<!-- /SYNC:double-round-trip-review --> <!-- SYNC:fresh-context-review -->Deep Multi-Round Review — Escalating rounds. Round 1 in main session. Round 2+ and EVERY recursive re-review iteration MUST use a fresh sub-agent.
Round 1: Main-session review. Read target files, build understanding, note issues. Output baseline findings.
Round 2: MANDATORY fresh sub-agent review — see
for the spawn mechanism andSYNC:fresh-context-reviewfor the canonical Agent prompt template. The sub-agent re-reads ALL files from scratch with ZERO Round 1 memory. It must catch:SYNC:review-protocol-injection
- Cross-cutting concerns missed in Round 1
- Interaction bugs between changed files
- Convention drift (new code vs existing patterns)
- Missing pieces that should exist but don't
- Subtle edge cases the main session rationalized away
Round 3+ (recursive after fixes): After ANY fix cycle, MANDATORY fresh sub-agent re-review. Spawn a NEW Agent tool call each iteration — never reuse Round 2's agent. Each new agent re-reads ALL files from scratch with full protocol injection. Continue until PASS or 3 fresh-subagent rounds max, then escalate to user via
.AskUserQuestionRules:
- NEVER declare PASS after Round 1 alone
- NEVER reuse a sub-agent across rounds — every iteration spawns a NEW Agent call
- Main agent READS sub-agent reports but MUST NOT filter, reinterpret, or override findings
- Max 3 fresh-subagent rounds per review — if still FAIL, escalate via
(do NOT silently loop)AskUserQuestion- Track round count in conversation context (session-scoped)
- Final verdict must incorporate ALL rounds
Report must include
for every round N≥2.## Round N Findings (Fresh Sub-Agent)
<!-- /SYNC:fresh-context-review --> <!-- SYNC:review-protocol-injection -->Fresh Sub-Agent Review — Eliminate orchestrator confirmation bias via isolated sub-agents.
Why: The main agent knows what it (or
) just fixed and rationalizes findings accordingly. A fresh sub-agent has ZERO memory, re-reads from scratch, and catches what the main agent dismissed. Sub-agent bias is mitigated by (1) fresh context, (2) verbatim protocol injection, (3) main agent not filtering the report./cookWhen: Round 2 of ANY review AND every recursive re-review iteration after fixes. NOT needed when Round 1 already PASSes with zero issues.
How:
- Spawn a NEW
tool call — useAgentsubagent_type for code reviews,code-reviewerfor plan/doc/artifact reviewsgeneral-purpose- Inject ALL required review protocols VERBATIM into the prompt — see
for the full list and template. Never reference protocols by file path; AI compliance drops behind file-read indirection (seeSYNC:review-protocol-injection)SYNC:shared-protocol-duplication-policy- Sub-agent re-reads ALL target files from scratch via its own tool calls — never pass file contents inline in the prompt
- Sub-agent writes structured report to
plans/reports/{review-type}-round{N}-{date}.md- Main agent reads the report, integrates findings into its own report, DOES NOT override or filter
Rules:
- NEVER reuse a sub-agent across rounds — every iteration spawns a NEW
callAgent- NEVER skip fresh-subagent review because "last round was clean" — every fix triggers a fresh round
- Max 3 fresh-subagent rounds per review — escalate via
if still failing; do NOT silently loop or fall back to any prior protocolAskUserQuestion- Track iteration count in conversation context (session-scoped, no persistent files)
Review Protocol Injection — Every fresh sub-agent review prompt MUST embed 10 protocol blocks VERBATIM. The template below has ALL 10 bodies already expanded inline. Copy the template wholesale into the Agent call's
field at runtime, replacing only thepromptin Task / Round / Reference Docs / Target Files / Output sections with context-specific values. Do NOT touch the embedded protocol sections.{placeholders}Why inline expansion: Placeholder markers would force file-read indirection at runtime. AI compliance drops significantly behind indirection (see
). Therefore the template carries all 10 protocol bodies pre-embedded.SYNC:shared-protocol-duplication-policy
Subagent Type Selection
— for code reviews (reviewing source files, git diffs, implementation)code-reviewer
— for plan / doc / artifact reviews (reviewing markdown plans, docs, specs)general-purpose
Canonical Agent Call Template (Copy Verbatim)
Agent({ description: "Fresh Round {N} review", subagent_type: "code-reviewer", prompt: ` ## Task {review-specific task — e.g., "Review all uncommitted changes for code quality" | "Review plan files under {plan-dir}" | "Review integration tests in {path}"} ## Round Round {N}. You have ZERO memory of prior rounds. Re-read all target files from scratch via your own tool calls. Do NOT trust anything from the main agent beyond this prompt. ## Protocols (follow VERBATIM — these are non-negotiable) ### Evidence-Based Reasoning Speculation is FORBIDDEN. Every claim needs proof. 1. Cite file:line, grep results, or framework docs for EVERY claim 2. Declare confidence: >80% act freely, 60-80% verify first, <60% DO NOT recommend 3. Cross-service validation required for architectural changes 4. "I don't have enough evidence" is valid and expected output BLOCKED until: Evidence file path (file:line) provided; Grep search performed; 3+ similar patterns found; Confidence level stated. Forbidden without proof: "obviously", "I think", "should be", "probably", "this is because". If incomplete → output: "Insufficient evidence. Verified: [...]. Not verified: [...]." ### Bug Detection MUST check categories 1-4 for EVERY review. Never skip. 1. Null Safety: Can params/returns be null? Are they guarded? Optional chaining gaps? .find() returns checked? 2. Boundary Conditions: Off-by-one (< vs <=)? Empty collections handled? Zero/negative values? Max limits? 3. Error Handling: Try-catch scope correct? Silent swallowed exceptions? Error types specific? Cleanup in finally? 4. Resource Management: Connections/streams closed? Subscriptions unsubscribed on destroy? Timers cleared? Memory bounded? 5. Concurrency (if async): Missing await? Race conditions on shared state? Stale closures? Retry storms? 6. Stack-Specific: JS: === vs ==, typeof null. C#: async void, missing using, LINQ deferred execution. Classify: CRITICAL (crash/corrupt) → FAIL | HIGH (incorrect behavior) → FAIL | MEDIUM (edge case) → WARN | LOW (defensive) → INFO. ### Design Patterns Quality Priority checks for every code change: 1. DRY via OOP: Same-suffix classes (*Entity, *Dto, *Service) MUST share base class. 3+ similar patterns → extract to shared abstraction. 2. Right Responsibility: Logic in LOWEST layer (Entity > Domain Service > Application Service > Controller). Never business logic in controllers. 3. SOLID: Single responsibility (one reason to change). Open-closed (extend, don't modify). Liskov (subtypes substitutable). Interface segregation (small interfaces). Dependency inversion (depend on abstractions). 4. After extraction/move/rename: Grep ENTIRE scope for dangling references. Zero tolerance. 5. YAGNI gate: NEVER recommend patterns unless 3+ occurrences exist. Don't extract for hypothetical future use. Anti-patterns to flag: God Object, Copy-Paste inheritance, Circular Dependency, Leaky Abstraction. ### Logic & Intention Review Verify WHAT code does matches WHY it was changed. 1. Change Intention Check: Every changed file MUST serve the stated purpose. Flag unrelated changes as scope creep. 2. Happy Path Trace: Walk through one complete success scenario through changed code. 3. Error Path Trace: Walk through one failure/edge case scenario through changed code. 4. Acceptance Mapping: If plan context available, map every acceptance criterion to a code change. NEVER mark review PASS without completing both traces (happy + error path). ### Test Spec Verification Map changed code to test specifications. 1. From changed files → find TC-{FEAT}-{NNN} in docs/business-features/{Service}/detailed-features/{Feature}.md Section 15. 2. Every changed code path MUST map to a corresponding TC (or flag as "needs TC"). 3. New functions/endpoints/handlers → flag for test spec creation. 4. Verify TC evidence fields point to actual code (file:line, not stale references). 5. Auth changes → TC-{FEAT}-02x exist? Data changes → TC-{FEAT}-01x exist? 6. If no specs exist → log gap and recommend /tdd-spec. NEVER skip test mapping. Untested code paths are the #1 source of production bugs. ### Fix-Layer Accountability NEVER fix at the crash site. Trace the full flow, fix at the owning layer. The crash site is a SYMPTOM, not the cause. MANDATORY before ANY fix: 1. Trace full data flow — Map the complete path from data origin to crash site across ALL layers (storage → backend → API → frontend → UI). Identify where bad state ENTERS, not where it CRASHES. 2. Identify the invariant owner — Which layer's contract guarantees this value is valid? Fix at the LOWEST layer that owns the invariant, not the highest layer that consumes it. 3. One fix, maximum protection — If fix requires touching 3+ files with defensive checks, you are at the wrong layer — go lower. 4. Verify no bypass paths — Confirm all data flows through the fix point. Check for direct construction skipping factories, clone/spread without re-validation, raw data not wrapped in domain models, mutations outside the model layer. BLOCKED until: Full data flow traced (origin → crash); Invariant owner identified with file:line evidence; All access sites audited (grep count); Fix layer justified (lowest layer that protects most consumers). Anti-patterns (REJECT): "Fix it where it crashes" (crash site ≠ cause site, trace upstream); "Add defensive checks at every consumer" (scattered defense = wrong layer); "Both fix is safer" (pick ONE authoritative layer). ### Rationalization Prevention AI skips steps via these evasions. Recognize and reject: - "Too simple for a plan" → Simple + wrong assumptions = wasted time. Plan anyway. - "I'll test after" → RED before GREEN. Write/verify test first. - "Already searched" → Show grep evidence with file:line. No proof = no search. - "Just do it" → Still need TaskCreate. Skip depth, never skip tracking. - "Just a small fix" → Small fix in wrong location cascades. Verify file:line first. - "Code is self-explanatory" → Future readers need evidence trail. Document anyway. - "Combine steps to save time" → Combined steps dilute focus. Each step has distinct purpose. ### Graph-Assisted Investigation MANDATORY when .code-graph/graph.db exists. HARD-GATE: MUST run at least ONE graph command on key files before concluding any investigation. Pattern: Grep finds files → trace --direction both reveals full system flow → Grep verifies details. - Investigation/Scout: trace --direction both on 2-3 entry files - Fix/Debug: callers_of on buggy function + tests_for - Feature/Enhancement: connections on files to be modified - Code Review: tests_for on changed functions - Blast Radius: trace --direction downstream CLI: python .claude/scripts/code_graph {command} --json. Use --node-mode file first (10-30x less noise), then --node-mode function for detail. ### Understand Code First HARD-GATE: Do NOT write, plan, or fix until you READ existing code. 1. Search 3+ similar patterns (grep/glob) — cite file:line evidence. 2. Read existing files in target area — understand structure, base classes, conventions. 3. Run python .claude/scripts/code_graph trace <file> --direction both --json when .code-graph/graph.db exists. 4. Map dependencies via connections or callers_of — know what depends on your target. 5. Write investigation to .ai/workspace/analysis/ for non-trivial tasks (3+ files). 6. Re-read analysis file before implementing — never work from memory alone. 7. NEVER invent new patterns when existing ones work — match exactly or document deviation. BLOCKED until: Read target files; Grep 3+ patterns; Graph trace (if graph.db exists); Assumptions verified with evidence. ## Reference Docs (READ before reviewing) - docs/project-reference/code-review-rules.md - {skill-specific reference docs — e.g., integration-test-reference.md for integration-test-review; backend-patterns-reference.md for backend reviews; frontend-patterns-reference.md for frontend reviews} ## Target Files {explicit file list OR "run git diff to see uncommitted changes" OR "read all files under {plan-dir}"} ## Output Write a structured report to plans/reports/{review-type}-round{N}-{date}.md with sections: - Status: PASS | FAIL - Issue Count: {number} - Critical Issues (with file:line evidence) - High Priority Issues (with file:line evidence) - Medium / Low Issues - Cross-cutting findings Return the report path and status to the main agent. Every finding MUST have file:line evidence. Speculation is forbidden. ` })
Rules
- DO copy the template wholesale — including all 10 embedded protocol sections
- DO replace only the
in Task / Round / Reference Docs / Target Files / Output sections with context-specific content{placeholders} - DO choose
subagent_type for code reviews andcode-reviewer
for plan / doc / artifact reviewsgeneral-purpose - DO NOT paraphrase, summarize, or skip any protocol section
- DO NOT pass file contents inline — the sub-agent reads via its own tool calls so it has a fresh context
- DO NOT reference protocols by file path or tag name — the bodies are already embedded above
- DO NOT introduce placeholder markers for the protocols — they must stay literally expanded
— Domain entity catalog, relationships, cross-service sync (read when task involves business entities/models) (content auto-injected by hook — check for [Injected: ...] header before reading)docs/project-reference/domain-entities-reference.md
— Test specifications by module (read existing TCs; generate/update test specs viadocs/test-specs/
after implementation)/tdd-spec
<!-- /SYNC:plan-quality -->Plan Quality — Every plan phase MUST ATTENTION include test specifications.
- Add
section with TC-{FEAT}-{NNN} IDs to every phase file## Test Specifications- Map every functional requirement to ≥1 TC (or explicit
with rationale)TBD- TC IDs follow
format — reference by ID, never embed full contentTC-{FEATURE}-{NNN}- Before any new workflow step: call
and re-read the phase fileTaskList- On context compaction: call
FIRST — never create duplicate tasksTaskList- Verify TC satisfaction per phase before marking complete (evidence must be
, not TBD)file:lineMode: TDD-first → reference existing TCs with
. Implement-first → use TBD →Evidence: TBDfills after./tdd-spec
<!-- SYNC:rationalization-prevention -->Evidence Gate: MANDATORY IMPORTANT MUST ATTENTION — every claim, finding, and recommendation requires
proof or traced evidence with confidence percentage (>80% to act, <80% must verify first).file:line
<!-- /SYNC:rationalization-prevention -->Rationalization Prevention — AI skips steps via these evasions. Recognize and reject:
Evasion Rebuttal "Too simple for a plan" Simple + wrong assumptions = wasted time. Plan anyway. "I'll test after" RED before GREEN. Write/verify test first. "Already searched" Show grep evidence with . No proof = no search.file:line"Just do it" Still need TaskCreate. Skip depth, never skip tracking. "Just a small fix" Small fix in wrong location cascades. Verify file:line first. "Code is self-explanatory" Future readers need evidence trail. Document anyway. "Combine steps to save time" Combined steps dilute focus. Each step has distinct purpose.
External Memory: For complex or lengthy work (research, analysis, scan, review), write intermediate findings and final results to a report file in
— prevents context loss and serves as deliverable.plans/reports/
Quick Summary
Goal: Implement a feature step-by-step with research, planning, execution, and verification.
Workflow:
- Question — Clarify requirements via AskUserQuestion; challenge assumptions
- Research — Use researcher subagents in parallel; scout codebase for patterns
- Plan — Create implementation plan, get user approval
- Implement — Execute with skill activation, code-simplifier, review-changes
Key Rules:
- Parent skill for all cook-* variants (cook-auto, cook-fast, cook-hard, cook-parallel)
- Write research findings to
for context preservation.ai/workspace/analysis/ - Always activate relevant skills from catalog during implementation
- Break work into small todo tasks; add final review task
Frontend/UI Context (if applicable)
<!-- SYNC:ui-system-context -->When this task involves frontend or UI changes,
<!-- /SYNC:ui-system-context -->UI System Context — For ANY task touching
,.ts,.html, or.scssfiles:.cssMUST ATTENTION READ before implementing:
— component base classes, stores, formsdocs/project-reference/frontend-patterns-reference.md — BEM methodology, SCSS variables, mixins, responsivedocs/project-reference/scss-styling-guide.md — design tokens, component inventory, iconsdocs/project-reference/design-system/README.mdReference
for project-specific paths.docs/project-config.json
- Component patterns:
docs/project-reference/frontend-patterns-reference.md - Styling/BEM guide:
docs/project-reference/scss-styling-guide.md - Design system tokens:
docs/project-reference/design-system/README.md
Pre-Implementation Granularity Gate (MANDATORY)
<HARD-GATE> <!-- SYNC:plan-granularity --><!-- /SYNC:plan-granularity -->Plan Granularity — Every phase must pass 5-point check before implementation:
- Lists exact file paths to modify (not generic "implement X")
- No planning verbs (research, investigate, analyze, determine, figure out)
- Steps ≤30min each, phase total ≤3h
- ≤5 files per phase
- No open decisions or TBDs in approach
Failing phases → create sub-plan. Repeat until ALL leaf phases pass (max depth: 3). Self-question: "Can I start coding RIGHT NOW? If any step needs 'figuring out' → sub-plan it."
If ANY check fails → STOP. Ask user: "Phase needs more detail before implementation. Refine with /plan? [Y/n]" DO NOT implement a phase that contains planning verbs, unnamed files, or unresolved decisions. </HARD-GATE>
Per-Phase Quality Cycle (MANDATORY)
<HARD-GATE> <!-- SYNC:iterative-phase-quality --><!-- /SYNC:iterative-phase-quality -->Iterative Phase Quality — Score complexity BEFORE planning.
Complexity signals: >5 files +2, cross-service +3, new pattern +2, DB migration +2 Score >=6 → MUST ATTENTION decompose into phases. Each phase:
- ≤5 files modified
- ≤3h effort
- Follows cycle: plan → implement → review → fix → verify
- Do NOT start Phase N+1 until Phase N passes VERIFY
Phase success = all TCs pass + code-reviewer agent approves + no CRITICAL findings.
Each plan phase = one quality cycle (plan→implement→review→fix→verify). DO NOT start next phase until current phase passes VERIFY. After each phase: re-assess remaining phases for scope changes. </HARD-GATE>
TC Satisfaction Verification (Per Phase)
After implementing each phase, before marking it complete:
- Read the phase's
section## Test Specifications - For each mapped TC: verify evidence exists (file:line, not TBD), grep-verify the file
- If any TC lacks evidence → phase is NOT complete
- Update phase file's TC table with actual evidence references
Greenfield Mode
Auto-detected when no code directories (
,src/,app/,lib/) or manifests (packages//package.json/*.sln) exist.go.mod
- Approved plan exists in
→ scaffold from planplans/ - No plan → redirect: "Run /plan first to create a greenfield project plan."
- Generate folder layout, starter files, build config, CI skeleton, CLAUDE.md
- Skip codebase pattern search. After scaffolding, run
./project-config
Variant Decision Guide
| If implementation needs... | Use | Why |
|---|---|---|
| Quick, straightforward | | Skip deep research, minimal planning |
| Complex, multi-layer | | Maximum verification, subagent research |
| Backend + frontend parallel | | Parallel fullstack-developer agents |
| Full autonomous execution | | Minimal user interaction |
| Fast autonomous | | Auto + skip deep research |
| Parallel autonomous | | Auto + parallel agents |
| General/interactive | (this skill) | Step-by-step with user collaboration |
Think harder to plan & start working on these tasks: <tasks>$ARGUMENTS</tasks>
Your Approach
- MUST ATTENTION use
to clarify — NEVER assume requirementsAskUserQuestion - MUST ATTENTION be brutally honest — flag unrealistic/over-engineered approaches directly
- MUST ATTENTION present 2-3 alternatives with pros/cons for non-trivial decisions
- MUST ATTENTION challenge initial approach — the best solution often differs from first instinct
Workflow
IMPORTANT: Analyze the skills catalog at
.claude/skills/* and activate needed skills during the process.
Research
- Parallel
subagents. Reports <=150 lines with citations.researcher
(preferred) or/scout-ext
(fallback) for codebase search./scout- MUST ATTENTION write findings to
. Re-read ENTIRE file before planning..ai/workspace/analysis/{task-name}.analysis.md
Plan
subagent with progressive disclosure:planner
(<=80 lines) +plan.md
per phase.phase-XX-name.md- Each phase: Context, Overview, Requirements, Architecture, Related Files, Steps, TCs, Success Criteria, Risks, Next Steps.
Implementation
to implement step by step./code
for product UIs./interface-design
for marketing/creative UIs./frontend-design
subagent for frontend perui-ux-designer
../docs/design-guidelines.md- MUST ATTENTION run type checking and compile after each change.
Subagent Discipline: Paste full task text (NEVER make subagent read plan file). Require "ask questions before starting". Require self-review before reporting.
Batch Checkpoint (Large Plans)
For plans with 10+ tasks, execute in batches with human review:
- Execute batch — Complete next 3 tasks (or user-specified batch size)
- Report — Show what was implemented, verification output, any concerns
- Wait — Say "Ready for feedback" and STOP. Do NOT continue automatically.
- Apply feedback — Incorporate changes, then execute next batch
- Repeat until all tasks complete
Testing
- Real tests: happy path, edge cases, error cases. NEVER fake data/mocks just to pass build.
subagent → failures →tester
subagent → fix → repeat until green.debugger- MANDATORY: After writing tests, MUST run them and verify they pass. If tests fail, diagnose root cause: (a) test code has wrong setup/assertions → fix test, or (b) service code has actual bug → report as finding. Never mark test task as done until tests actually pass.
Code Review
<!-- SYNC:two-stage-task-review --><!-- /SYNC:two-stage-task-review -->Two-Stage Task Review — Both stages MUST ATTENTION complete before marking task done.
Stage 1: Self-review — Immediately after implementation:
- Requirements met? No regressions? Code quality acceptable?
Stage 2: Cross-review — Via
subagent:code-reviewer
- Catches blind spots, convention drift, missed edge cases
NEVER skip Stage 2. Self-review alone misses 40%+ of issues.
(1)
spec-compliance-reviewer first, (2) code-reviewer after spec passes.
- Critical issues → fix → re-run
. Report summary to user for approval.tester
PM, Docs & Final Report
- Approved: Parallel
+project-manager
subagents. Rejected: Ask issues, fix, repeat.docs-manager - Final: summary of changes + next steps. Ask about commit/push via
.git-manager
Red Flags — STOP
| Evasion thought | Correct action |
|---|---|
| "Too simple for a plan" | Plan anyway. Hidden complexity. |
| "I already know how" | Check codebase patterns first. NEVER assume. |
| "Code first, test later" | Write test first. Or verify after EACH change. |
| "Plan is close enough" | Follow exactly or raise concerns. Drift compounds. |
| "Commit after everything" | Commit after each task. Frequent commits prevent loss. |
| "This refactor will improve things" | Only refactor what's in scope. YAGNI. |
| "Review is obvious, skip it" | NEVER skip. Reviews catch what authors miss. |
<!-- /SYNC:graph-assisted-investigation --> <!-- SYNC:incremental-persistence -->Graph-Assisted Investigation — MANDATORY when
exists..code-graph/graph.dbHARD-GATE: MUST ATTENTION run at least ONE graph command on key files before concluding any investigation.
Pattern: Grep finds files →
reveals full system flow → Grep verifies detailstrace --direction both
Task Minimum Graph Action Investigation/Scout on 2-3 entry filestrace --direction bothFix/Debug on buggy function +callers_oftests_forFeature/Enhancement on files to be modifiedconnectionsCode Review on changed functionstests_forBlast Radius trace --direction downstreamCLI:
. Usepython .claude/scripts/code_graph {command} --jsonfirst (10-30x less noise), then--node-mode filefor detail.--node-mode function
<!-- /SYNC:incremental-persistence -->Incremental Result Persistence — MANDATORY for all sub-agents or heavy inline steps processing >3 files.
- Before starting: Create report file
plans/reports/{skill}-{date}-{slug}.md- After each file/section reviewed: Append findings to report immediately — never hold in memory
- Return to main agent: Summary only (per SYNC:subagent-return-contract) with
pathFull report:- Main agent: Reads report file only when resolving specific blockers
Why: Context cutoff mid-execution loses ALL in-memory findings. Each disk write survives compaction. Partial results are better than no results.
Report naming:
plans/reports/{skill-name}-{YYMMDD}-{HHmm}-{slug}.md
After implementing, run
on modified files to verify no related files need updates.python .claude/scripts/code_graph connections <file> --json
Graph-Trace Before Implementation
MUST ATTENTION run BEFORE writing code when graph.db exists:
— callers + triggerspython .claude/scripts/code_graph trace <file-to-modify> --direction both --json
— all downstream consumerspython .claude/scripts/code_graph trace <file-to-modify> --direction downstream --json- Prevents breaking implicit dependencies (bus consumers, event handlers) invisible in the file itself.
Workflow Recommendation
MANDATORY IMPORTANT MUST ATTENTION — NO EXCEPTIONS: If you are NOT already in a workflow, you MUST ATTENTION use
to ask the user. Do NOT judge task complexity or decide this is "simple enough to skip" — the user decides whether to use a workflow, not you:AskUserQuestion
- Activate
workflow (Recommended) — scout → investigate → plan → cook → review → sre-review → test → docsfeature- Execute
directly — run this skill standalone/cook
Next Steps (Standalone: MUST ATTENTION ask user via AskUserQuestion
. Skip if inside workflow.)
AskUserQuestionMANDATORY IMPORTANT MUST ATTENTION — NO EXCEPTIONS after completing this skill, you MUST ATTENTION use
AskUserQuestion to present these options. Do NOT skip because the task seems "simple" or "obvious" — the user decides:
- "Proceed with full workflow (Recommended)" — I'll detect the best workflow to continue from here (feature implemented). This ensures review, testing, and docs steps aren't skipped.
- "/code-simplifier" — Simplify and clean up implementation
- "/integration-test" — Generate/update integration tests from test specs
- "/workflow-review-changes" — Review changes before commit
- "Skip, continue manually" — user decides
Standalone Review Gate (Non-Workflow Only)
MANDATORY IMPORTANT MUST ATTENTION: If this skill is called outside a workflow (standalone
), you MUST ATTENTION create a/cooktodo task forTaskCreateas the last task in your task list. This ensures all changes are reviewed before commit even without a workflow enforcing it./review-changesIf already running inside a workflow (e.g.,
,feature), skip this — the workflow sequence handlesbugfixat the appropriate step./review-changes
Closing Reminders
MANDATORY IMPORTANT MUST ATTENTION break work into small todo tasks using
TaskCreate BEFORE starting.
MANDATORY IMPORTANT MUST ATTENTION validate decisions with user via AskUserQuestion — never auto-decide.
MANDATORY IMPORTANT MUST ATTENTION add a final review todo task to verify work quality.
MANDATORY IMPORTANT MUST ATTENTION READ the following files before starting:
<!-- SYNC:plan-granularity:reminder -->
-
MANDATORY IMPORTANT MUST ATTENTION verify all phases pass 5-point granularity check. Failing phases → sub-plan. "Can I start coding RIGHT NOW?" <!-- /SYNC:plan-granularity:reminder -->
<!-- SYNC:understand-code-first:reminder --> -
MANDATORY IMPORTANT MUST ATTENTION search 3+ existing patterns and read code BEFORE any modification. Run graph trace when graph.db exists.
<!-- /SYNC:understand-code-first:reminder --> <!-- SYNC:plan-quality:reminder --> -
MANDATORY IMPORTANT MUST ATTENTION include
with TC IDs per phase. Call## Test Specifications
before creating new tasks. <!-- /SYNC:plan-quality:reminder --> <!-- SYNC:rationalization-prevention:reminder -->TaskList -
MANDATORY IMPORTANT MUST ATTENTION follow ALL steps regardless of perceived simplicity. "Too simple to plan" is an evasion, not a reason.
<!-- /SYNC:rationalization-prevention:reminder --> <!-- SYNC:ui-system-context:reminder --> -
MANDATORY IMPORTANT MUST ATTENTION read frontend-patterns-reference, scss-styling-guide, design-system/README before any UI change.
<!-- /SYNC:ui-system-context:reminder --> <!-- SYNC:iterative-phase-quality:reminder --> -
MANDATORY IMPORTANT MUST ATTENTION score complexity first. Score >=6 → decompose. Each phase: plan → implement → review → fix → verify. No skipping.
<!-- /SYNC:iterative-phase-quality:reminder --> <!-- SYNC:graph-assisted-investigation:reminder --> -
MANDATORY IMPORTANT MUST ATTENTION run at least ONE graph command on key files when graph.db exists. Pattern: grep → graph trace → grep verify.
<!-- /SYNC:graph-assisted-investigation:reminder --> <!-- SYNC:critical-thinking-mindset:reminder --> -
MUST ATTENTION apply critical thinking — every claim needs traced proof, confidence >80% to act. Anti-hallucination: never present guess as fact.
<!-- /SYNC:critical-thinking-mindset:reminder --> <!-- SYNC:ai-mistake-prevention:reminder --> -
MUST ATTENTION apply AI mistake prevention — holistic-first debugging, fix at responsible layer, surface ambiguity before coding, re-read files after compaction.
<!-- /SYNC:ai-mistake-prevention:reminder -->