Dotnet-skills dotnet-mcaf-human-review-planning
Apply MCAF human-review-planning guidance for a large AI-generated code drop by reading the target area, tracing the natural user and system flows, identifying the riskiest boundaries, and prioritizing the files a human should inspect first. Use when the codebase is too large to review line-by-line and you need a practical review sequence plus a prioritized file list.
install
source · Clone the upstream repo
git clone https://github.com/managedcode/dotnet-skills
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/managedcode/dotnet-skills "$T" && mkdir -p ~/.claude/skills && cp -r "$T/catalog/Platform/MCAF/skills/dotnet-mcaf-human-review-planning" ~/.claude/skills/managedcode-dotnet-skills-dotnet-mcaf-human-review-planning && rm -rf "$T"
manifest:
catalog/Platform/MCAF/skills/dotnet-mcaf-human-review-planning/SKILL.mdsource content
MCAF: Human Review Planning
Trigger On
- a large AI-generated code drop needs a human review plan
- the reviewer cannot inspect every line and needs prioritization
- the user asks which files are highest risk before doing manual review
- the user names a generated folder and wants a saved review plan for it
Value
- produce a concrete project delta: code, docs, config, tests, CI, or review artifact
- reduce ambiguity through explicit planning, verification, and final validation skills
- leave reusable project context so future tasks are faster and safer
Do Not Use For
- normal small pull-request review
- automated bug finding without creating a human review sequence
Inputs
- the target folder, feature area, or bounded context under review
- the main user journeys or operational flows involved
- any known architecture context, adjacent entities, or existing system rules
- any exact output path the user wants for the saved plan
Quick Start
- Read the nearest
and confirm scope and constraints.AGENTS.md - Run this skill's
through theWorkflow
until outcomes are acceptable.Ralph Loop - Return the
with concrete artifacts and verification evidence.Required Result Format
Workflow
- Read enough of the target area and its immediate boundaries to understand the generated code before planning review.
- Map the natural flow of operations first:
- sign up or authentication
- create
- update
- register or configure
- execute primary business action
- complete, archive, or finalize
- Use that flow to derive the most efficient human review sequence.
- Use the reviewer's domain knowledge as a force multiplier:
- compare the generated code against known architecture and existing entities
- look for places where the new feature should behave like nearby existing flows
- prioritize boundaries where generated code may drift from established system rules
- Identify high-risk review zones:
- entry points and orchestration layers
- persistence and state transitions
- cross-boundary integrations
- permissions, validation, and invariants
- side effects such as email, payments, jobs, or notifications
- Produce two separate outputs:
- prioritized review flow
- prioritized files or modules to inspect
- Present both outputs in chat.
- If the user asks for a durable artifact, save the plan to the exact docs path they requested; otherwise use
.docs/AREA/HUMAN_REVIEW_PLAN.md
Deliver
- a prioritized human review sequence
- a prioritized list of files or modules to inspect first
- both sections presented separately in chat
- a saved
when requestedHUMAN_REVIEW_PLAN.md
Validate
- the plan is grounded in actual code reading, not only the folder names
- the review order follows actual user or system flows
- high-risk files are explained, not only listed
- priorities account for likely mismatch against existing architecture or analogous entities
- the plan helps a human skip low-value line-by-line review
- the saved plan is readable without extra chat context
Ralph Loop
Use the Ralph Loop for every task, including docs, architecture, testing, and tooling work.
- Brainstorm first (mandatory):
- analyze current state
- define the problem, target outcome, constraints, and risks
- generate options and think through trade-offs before committing
- capture the recommended direction and open questions
- Plan second (mandatory):
- write a detailed execution plan from the chosen direction
- list final validation skills to run at the end, with order and reason
- Execute one planned step and produce a concrete delta.
- Review the result and capture findings with actionable next fixes.
- Apply fixes in small batches and rerun the relevant checks or review steps.
- Update the plan after each iteration.
- Repeat until outcomes are acceptable or only explicit exceptions remain.
- If a dependency is missing, bootstrap it or return
with explicit reason and fallback path.status: not_applicable
Required Result Format
:status
|complete
|clean
|improved
|configured
|not_applicableblocked
: concise plan and current iteration stepplan
: concrete changes madeactions_taken
: final skills run, or skipped with reasonsvalidation_skills
: commands, checks, or review evidence summaryverification
: top unresolved items orremainingnone
For setup-only requests with no execution, return
status: configured and exact next commands.
Load References
- read
for the output shapereferences/review-plan-format.md - read
when deciding what deserves human attention firstreferences/risk-signals.md
Example Requests
- "Plan a human review for this 40K-line AI-generated feature."
- "I cannot review every file. Tell me what to inspect first."
- "Trace the signup-to-completion flow and save a HUMAN_REVIEW_PLAN.md."
- "Look through the generated folder, give me two separate prioritized review lists, and save them under docs for this area."