Review-squad experts

Use when a project needs multi-perspective review — pre-launch audit, post-refactor check, inherited codebase assessment, or periodic health check. Dispatches parallel expert reviewer agents with persona framing.

install
source · Clone the upstream repo
git clone https://github.com/2389-research/review-squad
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/2389-research/review-squad "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/experts" ~/.claude/skills/2389-research-review-squad-experts && rm -rf "$T"
manifest: skills/experts/SKILL.md
source content

Expert Panel

Overview

Dispatch a panel of expert reviewer subagents in parallel to audit a project. Each agent adopts a specialist persona and reviews independently. Results are consolidated into a single severity-ranked report that feeds into an implementation plan.

When to Use

  • Pre-launch review ("ready to go live?")
  • Post-refactor audit ("did we break anything?")
  • Inherited codebase assessment ("what are we working with?")
  • Periodic health check
  • Any project type: web, mobile, API, CLI, infrastructure, data pipeline

Workflow

digraph expert_review {
  rankdir=TB;
  "User requests review" -> "Identify project type";
  "Identify project type" -> "Present default panel for that type";
  "Present default panel for that type" -> "Suggest stack-specific additions";
  "Suggest stack-specific additions" -> "Ask: add, remove, or adjust?";
  "Ask: add, remove, or adjust?" -> "Finalize panel";
  "Finalize panel" -> "Dispatch all agents in parallel (run_in_background)";
  "Dispatch all agents in parallel (run_in_background)" -> "As each completes, note key finding";
  "As each completes, note key finding" -> "All done?";
  "All done?" -> "As each completes, note key finding" [label="no"];
  "All done?" -> "Consolidate into severity-ranked table" [label="yes"];
  "Consolidate into severity-ranked table" -> "Present full report";
  "Present full report" -> "Offer to write implementation plan for fixes";
}

Default Panels by Project Type

Web (static sites, SPAs, server-rendered apps)

#ExpertFocus Areas
1SEO ExpertMeta tags, heading hierarchy, sitemap, robots.txt, URL structure, RSS, structured data
2Accessibility ExpertSemantic HTML, skip nav, ARIA, color contrast, keyboard nav, motion/animation
3Mobile UX ExpertViewport, responsive CSS, touch targets (44x44px min), font sizes, overflow
4Copy EditorSpelling, grammar, tone consistency across all templates and content
5Performance ExpertCSS/JS delivery, image optimization, fonts, caching, build config
6Security ReviewerHeaders, XSS vectors, sensitive data exposure, link security, CORS
7Social/Meta Tags SpecialistOpenGraph, Twitter cards, favicon, canonical URLs, share previews
8Web Standards ExpertHTML validation, correct element usage, spec compliance, ARIA misuse

API

#ExpertFocus Areas
1API Design ReviewerRESTful conventions, naming consistency, versioning, pagination, error responses
2Security ReviewerAuth/authz, input validation, rate limiting, OWASP API top 10
3Performance ReviewerQuery efficiency, N+1 problems, caching strategy, payload sizes
4Documentation ReviewerOpenAPI/Swagger completeness, example accuracy, error documentation
5Reliability ReviewerError handling, timeouts, retries, circuit breakers, graceful degradation
6Data Model ReviewerSchema design, migrations, indexes, constraints, data integrity

Mobile App (iOS/Android)

#ExpertFocus Areas
1UX ReviewerNavigation patterns, gesture handling, platform conventions, onboarding
2Accessibility ReviewerVoiceOver/TalkBack, dynamic type, contrast, touch targets
3Performance ReviewerStartup time, memory, battery, network efficiency, image handling
4Security ReviewerData storage, keychain/keystore usage, certificate pinning, auth flows
5Store Compliance ReviewerApp Store/Play Store guidelines, permissions justification, privacy labels
6Copy EditorUI text, error messages, onboarding copy, localization readiness

CLI Tool

#ExpertFocus Areas
1UX/Ergonomics ReviewerFlag naming, help text, output formatting, progressive disclosure
2Error Handling ReviewerError messages, exit codes, edge cases, graceful failures
3Compatibility ReviewerShell compatibility, OS support, path handling, encoding
4Security ReviewerInput sanitization, credential handling, file permissions
5Documentation ReviewerMan page / --help completeness, README, examples

Suggest project-specific additions. After presenting the default panel, analyze the tech stack and suggest reviewers that cover gaps. Examples:

  • E-commerce web app: Payment flow reviewer, product catalog reviewer
  • Next.js + Vercel: Deployment/config specialist
  • Auth-heavy app: Auth/session reviewer
  • i18n site/app: Localization reviewer
  • Data pipeline: Data quality reviewer, schema evolution reviewer

The user decides whether to add them — but always suggest what's relevant.

Agent Prompt Template

Every reviewer agent prompt MUST follow this structure:

You are a [ROLE] reviewing a [PROJECT TYPE] [before launch / after refactor / etc.].
Do NOT write any code — only research and report findings.
If browser MCP tools are available, use them to inspect the running site at [URL].

The project is at [PATH]. [Brief description with key tech details].

Review the following and report issues ranked by severity
(critical, important, minor):

1. [Area] — [What to check, where to look]
2. [Area] — [What to check, where to look]
...
10. [Area] — [What to check, where to look]

Check [key directories]. Report a prioritized list of findings.

Critical elements:

  • Persona first — "You are a [ROLE]" gives the agent expertise framing
  • No-code guard — "Do NOT write any code" prevents agents from fixing things
  • Browser MCP — If a dev server is running, include the URL so agents can inspect rendered output, not just source
  • Severity ranking — Forces structured output (critical/important/minor)
  • Numbered review areas — 8-12 specific areas per reviewer, tailored to their expertise
  • Directory hints — Tell them where to look (src/, config files, etc.)

Dispatch Pattern

Use the Agent tool with

run_in_background: true
for ALL reviewers. Dispatch all in a single message block for maximum parallelism.

Agent(description="SEO expert site review", subagent_type="general-purpose", run_in_background=true, prompt="...")
Agent(description="Accessibility expert review", subagent_type="general-purpose", run_in_background=true, prompt="...")
...all agents in one message block...

As each agent completes, briefly note the headline finding for the user. Wait until all are done before the full consolidation.

Consolidating Results

Compile all findings into a single table, deduplicated and ranked:

## [Review Type]: [Project Name]

### CRITICAL (must fix)

| # | Issue | Source |
|---|-------|--------|
| 1 | **[Issue description]** ([specific detail]) | [Which reviewer] |

### IMPORTANT (should fix soon)

| # | Issue | Source |
|---|-------|--------|

### MINOR (backlog)

| # | Issue | Source |
|---|-------|--------|

### DEFERRED (noted, not blocking)

| # | Issue | Source | Reason |
|---|-------|--------|--------|

Cross-referencing: When multiple reviewers flag the same issue, combine them into one row and list all sources (e.g.,

Standards, A11y, Security
).

After the Report

  1. Present the consolidated report
  2. Ask if the user wants to work through the fixes
  3. If yes, offer to write an implementation plan (use writing-plans skill if available)
  4. Tackle critical items first, then important, then minor

Common Mistakes

  • Not asking user to customize the panel — Always present the default panel and ask for changes before dispatching
  • Agents writing code — Without the "Do NOT write any code" guard, agents will start fixing things instead of reporting
  • Generic prompts — "Review the site for SEO" is too vague. List 8-12 specific areas with where to look
  • Sequential dispatch — All agents are independent. Always dispatch in parallel with
    run_in_background
  • No source attribution — The consolidated report must show which expert found each issue so the user can weigh credibility
  • Skipping the plan step — A list of findings without a plan to fix them is incomplete. Offer to create the implementation plan
  • Wrong panel for project type — Don't use the web panel for an API project. Match the default panel to the project type