Claude-skill-registry blind-review

Blind peer review for code - multiple agents critique independently, then synthesize privately to avoid groupthink

install
source · Clone the upstream repo
git clone https://github.com/majiayu000/claude-skill-registry
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/majiayu000/claude-skill-registry "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/data/blind-review" ~/.claude/skills/majiayu000-claude-skill-registry-blind-review && rm -rf "$T"
manifest: skills/data/blind-review/SKILL.md
source content

Blind Peer Review Skill

Based on: Harvard+Stanford+CMU "LLM Review" paper - blind peer review improves quality while isolation protects originality.

Key insight: When agents openly see each other's work, they copy and converge. Blind review + private synthesis prevents groupthink.

FULLY AUTONOMOUS EXECUTION

Claude executes ALL phases without user intervention. Never ask user to run commands.


Phase 1: Identify Review Target

Determine what to review:

  • If user specifies files/PR → use those
  • If recent changes exist → review uncommitted changes
  • If neither → ask user what to review
# Get files to review
git diff --name-only HEAD~1  # Recent commit
git diff --name-only         # Uncommitted changes

Phase 2: Parallel Independent Reviews (BLIND)

Launch 3 reviewer agents in parallel using a SINGLE message with multiple Task tool calls.

CRITICAL: Each agent works in ISOLATION - they do NOT see each other's prompts or outputs.

Reviewer 1: Architecture Strategist

Prompt: Review this code from an architectural perspective. Focus on:
- Component boundaries and responsibilities
- Coupling and cohesion
- Design pattern usage
- Scalability concerns

Do NOT suggest implementation details. Only identify structural issues.

Files to review: [FILE_LIST]

Reviewer 2: Pattern Recognition Specialist

Prompt: Analyze this code for patterns and anti-patterns. Focus on:
- Code duplication
- Naming consistency
- Common anti-patterns (god objects, shotgun surgery, etc.)
- Adherence to codebase conventions

Do NOT suggest architecture changes. Only identify pattern issues.

Files to review: [FILE_LIST]

Reviewer 3: Domain Expert (project-specific)

Prompt: Review this code for domain-specific correctness. For Random Timer:
- Timer logic accuracy
- Redux state management patterns
- React Native/Expo best practices
- Theme system usage

Do NOT suggest pattern changes. Only identify domain issues.

Files to review: [FILE_LIST]

Execution

// SINGLE message with 3 parallel Task calls
Task({
  subagent_type: "compound-engineering:review:architecture-strategist",
  prompt: "[Architecture review prompt with files]",
  description: "Blind architecture review"
})

Task({
  subagent_type: "compound-engineering:review:pattern-recognition-specialist",
  prompt: "[Pattern review prompt with files]",
  description: "Blind pattern review"
})

Task({
  subagent_type: "general-purpose",
  prompt: "[Domain review prompt with files]",
  description: "Blind domain review"
})

Phase 3: Collect Critiques (NO CROSS-POLLINATION)

Store each agent's output separately. Do NOT share outputs between phases.

critique_1 = [Architecture Strategist output]
critique_2 = [Pattern Recognition output]
critique_3 = [Domain Expert output]

Phase 4: Private Synthesis

A SINGLE synthesis agent receives all critiques and produces final recommendations.

Key: The synthesis agent sees critiques but reviewers never saw each other's work.

Prompt: You are synthesizing 3 independent code reviews into actionable improvements.

CRITIQUE 1 (Architecture):
{critique_1}

CRITIQUE 2 (Patterns):
{critique_2}

CRITIQUE 3 (Domain):
{critique_3}

Your task:
1. Identify overlapping concerns (high priority)
2. Identify unique insights from each reviewer
3. Resolve any contradictions by explaining trade-offs
4. Produce a prioritized list of improvements
5. For each improvement, provide specific file:line references

Output format:
## High Priority (multiple reviewers flagged)
- Issue: [description]
- Location: [file:line]
- Fix: [specific action]

## Medium Priority (single reviewer, significant)
...

## Low Priority (nice-to-have)
...

Phase 5: Execute Improvements (Optional)

If user wants fixes applied:

For each High Priority item:
1. Read the file
2. Make the specific edit
3. Mark as complete

For each Medium Priority item:
1. Ask user if they want it applied
2. If yes, execute

Output Format

After all phases complete, present:

## Blind Peer Review Complete

### Review Scope
- Files reviewed: [count]
- Reviewers: Architecture, Patterns, Domain
- Mode: Blind (isolated) + Private Synthesis

### High Priority Issues
[From synthesis]

### Medium Priority Issues
[From synthesis]

### Low Priority Issues
[From synthesis]

### Consensus Areas
Issues flagged by multiple reviewers:
- [List]

### Unique Insights
- Architecture: [unique finding]
- Patterns: [unique finding]
- Domain: [unique finding]

Why This Works

Traditional ReviewBlind Peer Review
Reviewers see each otherReviewers isolated
Ideas converge (groupthink)Ideas stay diverse
Later reviewers copy earlierAll reviews original
One perspective dominatesAll perspectives equal

The isolation phase protects originality. The synthesis phase captures quality feedback.


Automatic Invocation

Claude detects when blind review is appropriate:

  • After significant code changes (3+ files modified)
  • Before creating PRs with complex changes
  • When user mentions "review" in context of recent work

No user commands needed - Claude invokes this skill autonomously when beneficial.


ACT DON'T INSTRUCT

This skill is FULLY AUTONOMOUS:

  • Claude launches all agents directly
  • Claude collects and synthesizes results
  • Claude presents findings
  • Claude applies fixes if requested

NEVER tell user to run commands or invoke agents manually.