Claude-skill-registry audit-comprehensive

Run all 7 domain audits in staged waves and aggregate results

install
source · Clone the upstream repo
git clone https://github.com/majiayu000/claude-skill-registry
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/majiayu000/claude-skill-registry "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/data/audit-comprehensive" ~/.claude/skills/majiayu000-claude-skill-registry-audit-comprehensive && rm -rf "$T"
manifest: skills/data/audit-comprehensive/SKILL.md
source content

Comprehensive Multi-Domain Audit Orchestrator

Version: 2.1 (7-Domain Coverage with S0/S1 Escalation) Time Savings: 70% faster than sequential (175min → 50min) Stages: 3 stages with 4+3+1 agent configuration

What This Does: Spawns 7 specialized audit agents in staged waves (respecting max 4 concurrent limit), with verification checkpoints and S0/S1 escalation, then aggregates findings into a comprehensive report.


Overview

This skill orchestrates a complete codebase audit across all 7 domains:

  1. Code Quality (
    audit-code
    ) - Code hygiene, types, framework patterns
  2. Security (
    audit-security
    ) - Auth, input validation, OWASP compliance
  3. Performance (
    audit-performance
    ) - Load times, queries, caching
  4. Documentation (
    audit-documentation
    ) - README, API docs, architecture
  5. Refactoring (
    audit-refactoring
    ) - Technical debt, complexity, DRY
  6. Process/Automation (
    audit-process
    ) - CI/CD, testing, workflows
  7. Engineering Productivity (
    audit-engineering-productivity
    ) - DX, debugging, offline support

Output: Single unified report in

docs/audits/comprehensive/COMPREHENSIVE_AUDIT_REPORT.md


Execution Flow

┌─────────────────────────────────────────────────────┐
│ Pre-Flight Validation                               │
│   - Verify all 7 audit skills exist                 │
│   - Create output directory                         │
│   - Gather baselines (tests, lint, patterns)        │
│   - Load false positives database                   │
└─────────────────────────────────────────────────────┘
                          │
                          ▼
┌─────────────────────────────────────────────────────┐
│ Stage 1: Technical Core (4 agents parallel)         │
│   - audit-code                                      │
│   - audit-security                                  │
│   - audit-performance                               │
│   - audit-refactoring                               │
│                                                     │
│ Checkpoint:                                         │
│   ✓ Verify 4 report files exist and non-empty      │
│   ✓ S0/S1 Check: If security finds criticals →     │
│     notify user before proceeding                   │
└─────────────────────────────────────────────────────┘
                          │
                          ▼
┌─────────────────────────────────────────────────────┐
│ Stage 2: Supporting (3 agents parallel)             │
│   - audit-documentation                             │
│   - audit-process                                   │
│   - audit-engineering-productivity                  │
│                                                     │
│ Checkpoint:                                         │
│   ✓ Verify 3 report files exist and non-empty      │
└─────────────────────────────────────────────────────┘
                          │
                          ▼
┌─────────────────────────────────────────────────────┐
│ Stage 3: Aggregation (sequential)                   │
│   - audit-aggregator                                │
│                                                     │
│ Checkpoint:                                         │
│   ✓ Verify COMPREHENSIVE_AUDIT_REPORT.md exists    │
└─────────────────────────────────────────────────────┘
                          │
                          ▼
┌─────────────────────────────────────────────────────┐
│ Post-Audit                                          │
│   - Update AUDIT_TRACKER.md                         │
│   - Display final summary                           │
│   - Recommend next steps                            │
└─────────────────────────────────────────────────────┘

Pre-Flight Validation

Step 0: Episodic Memory Search (Session #128)

Before running audits, search for context from past audit sessions:

// Search for past comprehensive audits
mcp__plugin_episodic -
  memory_episodic -
  memory__search({
    query: ["comprehensive audit", "findings", "patterns"],
    limit: 5,
  });

// Search for specific domain context
mcp__plugin_episodic -
  memory_episodic -
  memory__search({
    query: ["security audit", "S0", "critical"],
    limit: 5,
  });

Why this matters:

  • Compare against previous audit findings
  • Identify recurring issues (may indicate architectural debt)
  • Avoid flagging known false positives
  • Track improvement/regression trends

Use findings to:

  1. Note which S0/S1 issues from past audits are still open
  2. Identify patterns that keep appearing (root cause needed)
  3. Set context for aggregator on known false positives

Step 1: Verify Skills Exist

Check that all 7 audit skills are available:

ls -1 .claude/skills/audit-*/SKILL.md | wc -l
# Should return 7

If not all present, notify user which audits are missing and ask whether to proceed with available audits only.

Step 2: Create Output Directory

mkdir -p docs/audits/comprehensive

Step 2.5: Verify Output Directory (CRITICAL)

Before running ANY agent, verify AUDIT_DIR is valid:

AUDIT_DIR="docs/audits/comprehensive"
AUDIT_PATH=$(realpath "${AUDIT_DIR}" 2>/dev/null || echo "${AUDIT_DIR}")
# Check for empty, root path, or path traversal attempts
if [ -z "${AUDIT_DIR}" ] || [ "${AUDIT_PATH}" = "/" ] || [[ "${AUDIT_DIR}" == ".."* ]]; then
  echo "FATAL: Invalid or unsafe AUDIT_DIR"
  exit 1
fi
echo "Output directory: ${AUDIT_DIR}"

Why: Context compaction can cause variable loss. Always verify before agent launches.

Step 3: Run Baseline Checks

Gather current metrics to provide to all audits:

# Test count
npm test 2>&1 | grep -E "Tests:|passing|failed" | head -5

# Lint status
npm run lint 2>&1 | tail -10

# Pattern compliance
npm run patterns:check 2>&1 | head -20

Store results in

docs/audits/comprehensive/baseline.txt
for reference.

Step 4: Load False Positives

Read

docs/audits/FALSE_POSITIVES.jsonl
to pass to aggregator (prevents duplicate flagging of known false positives).


Stage 1: Technical Core Audits (4 Parallel)

Launch 4 agents IN PARALLEL using Task tool with

run_in_background: true
:

AgentSkillOutput File
1Aaudit-code
audit-code-report.md
1Baudit-security
audit-security-report.md
1Caudit-performance
audit-performance-report.md
1Daudit-refactoring
audit-refactoring-report.md

Why these 4 first:

  • Core technical analysis
  • Security findings needed for S0/S1 escalation check
  • Respects max 4 concurrent agents (CLAUDE.md Section 6.3)

Display Initial Status:

🚀 Comprehensive Audit Started
━━━━━━━━━━━━━━━━━━━━━━━━━━━━

Stage 1: Technical Core (4 parallel)
  ⏳ Code Quality
  ⏳ Security
  ⏳ Performance
  ⏳ Refactoring

Stage 2: Supporting (waiting)
  ⏸️ Documentation
  ⏸️ Process/Automation
  ⏸️ Engineering Productivity

Stage 3: Aggregation (waiting)
  ⏸️ Aggregator

Estimated time: 45-50 minutes
(vs 175 minutes if run sequentially - 70% faster!)

Stage 1 Checkpoint (MANDATORY)

Before proceeding to Stage 2, perform these checks:

1. Verify output files exist:

for f in audit-code-report.md audit-security-report.md audit-performance-report.md audit-refactoring-report.md; do
  if [ ! -s "docs/audits/comprehensive/$f" ]; then
    echo "❌ MISSING: $f - re-run agent"
  else
    echo "✅ $f exists"
  fi
done

2. S0/S1 Security Escalation Check:

grep -cE "\bS0\b|\bS1\b" docs/audits/comprehensive/audit-security-report.md

If S0/S1 findings exist, display:

⚠️ SECURITY ESCALATION
━━━━━━━━━━━━━━━━━━━━━━━

Security audit found critical/high findings.
These should be reviewed before continuing.

S0 Critical: X findings
S1 High: Y findings

Options:
1. Review security findings now (recommended for S0)
2. Continue with remaining audits
3. Stop and address security issues first

What would you like to do?

3. Display Stage 1 Summary:

✅ Stage 1 Complete (Technical Core)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━

  ✅ Code Quality    (X findings)
  ✅ Security        (X findings, Y critical)
  ✅ Performance     (X findings)
  ✅ Refactoring     (X findings)

Proceeding to Stage 2...

Stage 2: Supporting Audits (3 Parallel)

Launch 3 agents IN PARALLEL using Task tool with

run_in_background: true
:

AgentSkillOutput File
2Aaudit-documentation
audit-documentation-report.md
2Baudit-process
audit-process-report.md
2Caudit-engineering-productivity
audit-engineering-productivity-report.md

Why these in Stage 2:

  • Supporting audits that can use Stage 1 context
  • Lower priority than technical core
  • Completes the full 7-domain coverage

Stage 2 Checkpoint (MANDATORY)

1. Verify output files exist:

for f in audit-documentation-report.md audit-process-report.md audit-engineering-productivity-report.md; do
  if [ ! -s "docs/audits/comprehensive/$f" ]; then
    echo "❌ MISSING: $f - re-run agent"
  else
    echo "✅ $f exists"
  fi
done

2. Display Stage 2 Summary:

✅ Stage 2 Complete (Supporting)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━

  ✅ Documentation         (X findings)
  ✅ Process/Auto          (X findings)
  ✅ Engineering Productivity (X findings)

All 7 audits complete. Proceeding to aggregation...

Stage 3: Aggregation Phase

Launch Aggregator Agent

Use Task tool to spawn

audit-aggregator
agent:

Task({
  subagent_type: "audit-aggregator",
  description: "Aggregate and deduplicate audit results",
  prompt: `
Read all 7 audit reports from docs/audits/comprehensive/

Perform:
1. Deduplicate findings (same file:line across multiple audits → merge)
2. Identify cross-cutting patterns (files appearing in 3+ audits)
3. Priority ranking (severity × cross-domain count × effort)
4. Generate executive summary with top 20 findings

Output to: docs/audits/comprehensive/COMPREHENSIVE_AUDIT_REPORT.md
  `,
});

Expected Output:

  • docs/audits/comprehensive/COMPREHENSIVE_AUDIT_REPORT.md
    (unified report)

Wait for aggregator to complete (typically 3-5 minutes)

Stage 3 Checkpoint (MANDATORY)

After aggregator completes, verify:

if [ ! -s "docs/audits/comprehensive/COMPREHENSIVE_AUDIT_REPORT.md" ]; then
  echo "❌ Aggregation failed - report not generated"
  echo "Individual reports still available for manual review"
else
  echo "✅ Comprehensive report generated"
fi

Final Report Structure

The

COMPREHENSIVE_AUDIT_REPORT.md
should contain:

Executive Summary

  • Total unique findings (after deduplication)
  • Severity breakdown (S0: X, S1: Y, S2: Z, S3: W)
  • Top 3 cross-domain insights
  • Recommended fix order
  • Effort estimate (total hours)

Priority-Ranked Findings (Top 20)

Table format:

RankIDSeverityDomainsFile:LineDescriptionEffort
1COMP-001S03auth.ts:45Missing auth checkE1
.....................

Cross-Domain Insights

Examples:

  • "Files X, Y, Z appear in 4+ audits → Comprehensive refactor needed"
  • "Security + Performance overlap: 12 findings where fixing one helps both"
  • "Documentation gaps align with code complexity hotspots"

Full Findings (Deduplicated)

Complete table of all findings grouped by severity, with links to original audit reports.

Appendix

  • Links to individual audit reports
  • Baseline metrics snapshot
  • False positives excluded (count)

Post-Audit (MANDATORY)

After aggregation completes, you MUST update tracking:

1. Update AUDIT_TRACKER.md

Add an entry to each of the 6 category tables in

docs/AUDIT_TRACKER.md
:

DateSessionCommits CoveredFiles CoveredFindingsReset Threshold
{TODAY}ComprehensiveFull codebaseAllSession #{N} - [report link]✅ (all)

2. Update Threshold Summary Table

In the "Current Thresholds" section, update all 6 categories:

  • Set "Last Audit" to today's date with "(Comprehensive)"
  • Reset "Commits Since" to 0
  • Reset "Files Since" to 0

3. Update Multi-AI Thresholds

In the "Multi-AI Audit Thresholds" section:

  • Update "Total commits" reset date to today
  • Update "Time elapsed" to "0 days (comprehensive audit {TODAY})"

This step ensures

npm run review:check
correctly shows no triggers after the audit.


Triage & Roadmap Integration (MANDATORY)

After TDMS intake completes, triage new items into the roadmap:

1. Review New Items

Check the newly added DEBT-XXXX items:

# View recent additions (last 50 items by ID)
tail -50 docs/technical-debt/MASTER_DEBT.jsonl | jq -r '[.id, .severity, .category, .title[:60]] | @tsv'

2. Priority Scoring

Beyond S0-S3 severity, consider these factors for prioritization:

FactorWeightDescription
Severity40%S0=100, S1=50, S2=20, S3=5
Cross-domain20%Items flagged by multiple audits get +50%
Effort inverse20%E0=4x, E1=2x, E2=1x, E3=0.5x (quick wins first)
Dependency10%Blockers for other items get +25%
File hotspot10%Files with 3+ findings get +25%

Priority Score Formula:

score = (severity × 0.4) × (cross_domain_mult × 0.2) × (effort_inv × 0.2) × (dep_mult × 0.1) × (hotspot_mult × 0.1)

3. Track Assignment

New items are auto-assigned based on category + file patterns:

CategoryFile PatternTrack
security*Track-S
performance*Track-P
process*Track-D
refactoring*M2.3-REF
documentation*M1.5
code-qualityscripts/, .claude/Track-E
code-quality.github/Track-D
code-qualitytests/Track-T
code-qualityfunctions/M2.2
code-qualitycomponents/, lib/, app/M2.1

View current assignments:

cat docs/technical-debt/views/unplaced-items.md

4. Update ROADMAP.md

For S0/S1 items that need immediate attention:

## Track-S: Security Technical Debt

- [ ] DEBT-0875: Firebase credentials written to disk (S1) **NEW**
- [ ] DEBT-0876: Missing App Check validation (S1) **NEW**

For bulk items by track:

- [ ] DEBT-0869 through DEBT-0880: Process automation gaps (S2, bulk)

5. Consistency Check

Verify all references are valid:

node scripts/debt/sync-roadmap-refs.js --check-only

Reports:

  • Orphaned refs (in ROADMAP but not in MASTER_DEBT)
  • Unplaced items (in MASTER_DEBT but not in ROADMAP)
  • Status mismatches (marked done but not RESOLVED)

6. Review Cadence

TriggerAction
After comprehensive auditFull triage of all new items
After single-domain auditTriage items in that category only
Weekly (if no audits)Check unplaced-items.md for drift
Before sprint planningReview S0/S1 items for inclusion

Completion

Display Final Summary:

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
🎯 COMPREHENSIVE AUDIT COMPLETE
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━

📊 Results Summary:
   • 142 raw findings → 97 unique (45 merged)
   • S0 Critical: 3
   • S1 High: 24
   • S2 Medium: 42
   • S3 Low: 28

🔍 Cross-Domain Insights:
   • 8 files need comprehensive refactor (4+ audits)
   • 12 security/performance overlaps
   • 5 documentation gaps in complex areas

📄 Full Report:
   docs/audits/comprehensive/COMPREHENSIVE_AUDIT_REPORT.md

⏱️  Total Time: ~45 minutes
   (vs 150 minutes sequential - saved ~105 minutes!)

🎯 Recommended Next Steps:
   1. Review top 20 priority findings
   2. Create GitHub issues for S0/S1
   3. Plan refactor for hotspot files

Error Handling

If Individual Audit Fails:

  • Continue with remaining audits in the same stage
  • Mark failed audit in status display (❌)
  • Note failure in final report
  • Suggest running failed audit individually for debugging

If Aggregator Fails:

  • All individual reports still available
  • User can manually review 6 separate reports
  • Suggest creating GitHub issue for aggregator failure

If All Audits Fail:

  • Check baseline environment (tests passing, lint working)
  • Check for system issues (disk space, memory)
  • Suggest running single audit first to isolate issue

Context Recovery

If context compacts mid-audit, resume from last completed checkpoint:

Determine Current State

echo "=== Checking audit progress ==="
ls -la docs/audits/comprehensive/*.md 2>/dev/null | wc -l

Recovery Matrix

Files FoundStateResume Action
0-3 reportsStage 1 incompleteRe-run missing Stage 1 audits
4 reportsStage 1 completeStart Stage 2
5 reportsStage 2 incompleteRe-run missing Stage 2 audit
6 reports, no COMPREHENSIVEStage 2 completeRun Stage 3 (aggregator)
COMPREHENSIVE existsCompleteRun post-audit only

Resume Commands

Stage 1 incomplete: Re-run only missing audits:

# Check which are missing
for audit in code security performance refactoring; do
  [ ! -f "docs/audits/comprehensive/audit-${audit}-report.md" ] && echo "Missing: $audit"
done

Stage 2 incomplete: Run documentation and/or process audits as needed.

Stage 3: Run aggregator on existing 6 reports.


Usage Examples

Quarterly Health Check:

/audit-comprehensive

Pre-Release Audit:

/audit-comprehensive

After Major Refactor:

/audit-comprehensive

Focused Audit (Not Comprehensive):

Use individual skills instead:

/audit-security   (25 min - when you only need security review)
/audit-code       (30 min - when you only need code quality)

Notes

  • Staged Execution: 3 stages (4+2+1 agents) respects CLAUDE.md max 4 concurrent limit
  • Time Estimate: ~45 minutes (vs 150min sequential = 70% savings)
  • S0/S1 Escalation: Security findings checked after Stage 1 before proceeding
  • Checkpoints: Each stage verifies outputs exist and are non-empty
  • Context Recovery: Can resume from any checkpoint after context compaction
  • Output Consistency: All audits use same severity (S0-S3) and effort (E0-E3) scales

Future Enhancements

  • Incremental Audits: Only re-run audits for changed domains
  • Custom Audit Subset:
    --audits code,security
    to run subset
  • Confidence Scoring: Weight findings by audit confidence levels
  • Trend Analysis: Compare against previous comprehensive audits
  • Auto-Issue Creation: Create GitHub issues for S0/S1 findings automatically

Related Skills

  • /audit-code
    - Individual code quality audit
  • /audit-security
    - Individual security audit
  • /audit-performance
    - Individual performance audit
  • /audit-documentation
    - Individual documentation audit
  • /audit-refactoring
    - Individual refactoring audit
  • /audit-process
    - Individual process/automation audit
  • /audit-aggregator
    - Standalone aggregation (if you have existing reports)

Documentation References

Before running this audit, review:

TDMS Integration (Required)

  • PROCEDURE.md - Full TDMS workflow
  • MASTER_DEBT.jsonl - Canonical debt store
  • All individual audits automatically run TDMS intake after completion

Documentation Standards (Required)


Version History

VersionDateDescription
2.12026-02-03Added Triage & Roadmap Integration section with priority scoring formula
2.02026-02-02Staged execution (4+2+1), S0/S1 escalation, checkpoints, context recovery
1.02026-01-28Initial version - flat parallel execution of all 6 audits