Claude-skill-registry audit-comprehensive
Run all 7 domain audits in staged waves and aggregate results
git clone https://github.com/majiayu000/claude-skill-registry
T=$(mktemp -d) && git clone --depth=1 https://github.com/majiayu000/claude-skill-registry "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/data/audit-comprehensive" ~/.claude/skills/majiayu000-claude-skill-registry-audit-comprehensive && rm -rf "$T"
skills/data/audit-comprehensive/SKILL.mdComprehensive Multi-Domain Audit Orchestrator
Version: 2.1 (7-Domain Coverage with S0/S1 Escalation) Time Savings: 70% faster than sequential (175min → 50min) Stages: 3 stages with 4+3+1 agent configuration
What This Does: Spawns 7 specialized audit agents in staged waves (respecting max 4 concurrent limit), with verification checkpoints and S0/S1 escalation, then aggregates findings into a comprehensive report.
Overview
This skill orchestrates a complete codebase audit across all 7 domains:
- Code Quality (
) - Code hygiene, types, framework patternsaudit-code - Security (
) - Auth, input validation, OWASP complianceaudit-security - Performance (
) - Load times, queries, cachingaudit-performance - Documentation (
) - README, API docs, architectureaudit-documentation - Refactoring (
) - Technical debt, complexity, DRYaudit-refactoring - Process/Automation (
) - CI/CD, testing, workflowsaudit-process - Engineering Productivity (
) - DX, debugging, offline supportaudit-engineering-productivity
Output: Single unified report in
docs/audits/comprehensive/COMPREHENSIVE_AUDIT_REPORT.md
Execution Flow
┌─────────────────────────────────────────────────────┐ │ Pre-Flight Validation │ │ - Verify all 7 audit skills exist │ │ - Create output directory │ │ - Gather baselines (tests, lint, patterns) │ │ - Load false positives database │ └─────────────────────────────────────────────────────┘ │ ▼ ┌─────────────────────────────────────────────────────┐ │ Stage 1: Technical Core (4 agents parallel) │ │ - audit-code │ │ - audit-security │ │ - audit-performance │ │ - audit-refactoring │ │ │ │ Checkpoint: │ │ ✓ Verify 4 report files exist and non-empty │ │ ✓ S0/S1 Check: If security finds criticals → │ │ notify user before proceeding │ └─────────────────────────────────────────────────────┘ │ ▼ ┌─────────────────────────────────────────────────────┐ │ Stage 2: Supporting (3 agents parallel) │ │ - audit-documentation │ │ - audit-process │ │ - audit-engineering-productivity │ │ │ │ Checkpoint: │ │ ✓ Verify 3 report files exist and non-empty │ └─────────────────────────────────────────────────────┘ │ ▼ ┌─────────────────────────────────────────────────────┐ │ Stage 3: Aggregation (sequential) │ │ - audit-aggregator │ │ │ │ Checkpoint: │ │ ✓ Verify COMPREHENSIVE_AUDIT_REPORT.md exists │ └─────────────────────────────────────────────────────┘ │ ▼ ┌─────────────────────────────────────────────────────┐ │ Post-Audit │ │ - Update AUDIT_TRACKER.md │ │ - Display final summary │ │ - Recommend next steps │ └─────────────────────────────────────────────────────┘
Pre-Flight Validation
Step 0: Episodic Memory Search (Session #128)
Before running audits, search for context from past audit sessions:
// Search for past comprehensive audits mcp__plugin_episodic - memory_episodic - memory__search({ query: ["comprehensive audit", "findings", "patterns"], limit: 5, }); // Search for specific domain context mcp__plugin_episodic - memory_episodic - memory__search({ query: ["security audit", "S0", "critical"], limit: 5, });
Why this matters:
- Compare against previous audit findings
- Identify recurring issues (may indicate architectural debt)
- Avoid flagging known false positives
- Track improvement/regression trends
Use findings to:
- Note which S0/S1 issues from past audits are still open
- Identify patterns that keep appearing (root cause needed)
- Set context for aggregator on known false positives
Step 1: Verify Skills Exist
Check that all 7 audit skills are available:
ls -1 .claude/skills/audit-*/SKILL.md | wc -l # Should return 7
If not all present, notify user which audits are missing and ask whether to proceed with available audits only.
Step 2: Create Output Directory
mkdir -p docs/audits/comprehensive
Step 2.5: Verify Output Directory (CRITICAL)
Before running ANY agent, verify AUDIT_DIR is valid:
AUDIT_DIR="docs/audits/comprehensive" AUDIT_PATH=$(realpath "${AUDIT_DIR}" 2>/dev/null || echo "${AUDIT_DIR}") # Check for empty, root path, or path traversal attempts if [ -z "${AUDIT_DIR}" ] || [ "${AUDIT_PATH}" = "/" ] || [[ "${AUDIT_DIR}" == ".."* ]]; then echo "FATAL: Invalid or unsafe AUDIT_DIR" exit 1 fi echo "Output directory: ${AUDIT_DIR}"
Why: Context compaction can cause variable loss. Always verify before agent launches.
Step 3: Run Baseline Checks
Gather current metrics to provide to all audits:
# Test count npm test 2>&1 | grep -E "Tests:|passing|failed" | head -5 # Lint status npm run lint 2>&1 | tail -10 # Pattern compliance npm run patterns:check 2>&1 | head -20
Store results in
docs/audits/comprehensive/baseline.txt for reference.
Step 4: Load False Positives
Read
docs/audits/FALSE_POSITIVES.jsonl to pass to aggregator (prevents
duplicate flagging of known false positives).
Stage 1: Technical Core Audits (4 Parallel)
Launch 4 agents IN PARALLEL using Task tool with
:run_in_background: true
| Agent | Skill | Output File |
|---|---|---|
| 1A | audit-code | |
| 1B | audit-security | |
| 1C | audit-performance | |
| 1D | audit-refactoring | |
Why these 4 first:
- Core technical analysis
- Security findings needed for S0/S1 escalation check
- Respects max 4 concurrent agents (CLAUDE.md Section 6.3)
Display Initial Status:
🚀 Comprehensive Audit Started ━━━━━━━━━━━━━━━━━━━━━━━━━━━━ Stage 1: Technical Core (4 parallel) ⏳ Code Quality ⏳ Security ⏳ Performance ⏳ Refactoring Stage 2: Supporting (waiting) ⏸️ Documentation ⏸️ Process/Automation ⏸️ Engineering Productivity Stage 3: Aggregation (waiting) ⏸️ Aggregator Estimated time: 45-50 minutes (vs 175 minutes if run sequentially - 70% faster!)
Stage 1 Checkpoint (MANDATORY)
Before proceeding to Stage 2, perform these checks:
1. Verify output files exist:
for f in audit-code-report.md audit-security-report.md audit-performance-report.md audit-refactoring-report.md; do if [ ! -s "docs/audits/comprehensive/$f" ]; then echo "❌ MISSING: $f - re-run agent" else echo "✅ $f exists" fi done
2. S0/S1 Security Escalation Check:
grep -cE "\bS0\b|\bS1\b" docs/audits/comprehensive/audit-security-report.md
If S0/S1 findings exist, display:
⚠️ SECURITY ESCALATION ━━━━━━━━━━━━━━━━━━━━━━━ Security audit found critical/high findings. These should be reviewed before continuing. S0 Critical: X findings S1 High: Y findings Options: 1. Review security findings now (recommended for S0) 2. Continue with remaining audits 3. Stop and address security issues first What would you like to do?
3. Display Stage 1 Summary:
✅ Stage 1 Complete (Technical Core) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ ✅ Code Quality (X findings) ✅ Security (X findings, Y critical) ✅ Performance (X findings) ✅ Refactoring (X findings) Proceeding to Stage 2...
Stage 2: Supporting Audits (3 Parallel)
Launch 3 agents IN PARALLEL using Task tool with
:run_in_background: true
| Agent | Skill | Output File |
|---|---|---|
| 2A | audit-documentation | |
| 2B | audit-process | |
| 2C | audit-engineering-productivity | |
Why these in Stage 2:
- Supporting audits that can use Stage 1 context
- Lower priority than technical core
- Completes the full 7-domain coverage
Stage 2 Checkpoint (MANDATORY)
1. Verify output files exist:
for f in audit-documentation-report.md audit-process-report.md audit-engineering-productivity-report.md; do if [ ! -s "docs/audits/comprehensive/$f" ]; then echo "❌ MISSING: $f - re-run agent" else echo "✅ $f exists" fi done
2. Display Stage 2 Summary:
✅ Stage 2 Complete (Supporting) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ ✅ Documentation (X findings) ✅ Process/Auto (X findings) ✅ Engineering Productivity (X findings) All 7 audits complete. Proceeding to aggregation...
Stage 3: Aggregation Phase
Launch Aggregator Agent
Use Task tool to spawn
audit-aggregator agent:
Task({ subagent_type: "audit-aggregator", description: "Aggregate and deduplicate audit results", prompt: ` Read all 7 audit reports from docs/audits/comprehensive/ Perform: 1. Deduplicate findings (same file:line across multiple audits → merge) 2. Identify cross-cutting patterns (files appearing in 3+ audits) 3. Priority ranking (severity × cross-domain count × effort) 4. Generate executive summary with top 20 findings Output to: docs/audits/comprehensive/COMPREHENSIVE_AUDIT_REPORT.md `, });
Expected Output:
(unified report)docs/audits/comprehensive/COMPREHENSIVE_AUDIT_REPORT.md
Wait for aggregator to complete (typically 3-5 minutes)
Stage 3 Checkpoint (MANDATORY)
After aggregator completes, verify:
if [ ! -s "docs/audits/comprehensive/COMPREHENSIVE_AUDIT_REPORT.md" ]; then echo "❌ Aggregation failed - report not generated" echo "Individual reports still available for manual review" else echo "✅ Comprehensive report generated" fi
Final Report Structure
The
COMPREHENSIVE_AUDIT_REPORT.md should contain:
Executive Summary
- Total unique findings (after deduplication)
- Severity breakdown (S0: X, S1: Y, S2: Z, S3: W)
- Top 3 cross-domain insights
- Recommended fix order
- Effort estimate (total hours)
Priority-Ranked Findings (Top 20)
Table format:
| Rank | ID | Severity | Domains | File:Line | Description | Effort |
|---|---|---|---|---|---|---|
| 1 | COMP-001 | S0 | 3 | auth.ts:45 | Missing auth check | E1 |
| ... | ... | ... | ... | ... | ... | ... |
Cross-Domain Insights
Examples:
- "Files X, Y, Z appear in 4+ audits → Comprehensive refactor needed"
- "Security + Performance overlap: 12 findings where fixing one helps both"
- "Documentation gaps align with code complexity hotspots"
Full Findings (Deduplicated)
Complete table of all findings grouped by severity, with links to original audit reports.
Appendix
- Links to individual audit reports
- Baseline metrics snapshot
- False positives excluded (count)
Post-Audit (MANDATORY)
After aggregation completes, you MUST update tracking:
1. Update AUDIT_TRACKER.md
Add an entry to each of the 6 category tables in
docs/AUDIT_TRACKER.md:
| Date | Session | Commits Covered | Files Covered | Findings | Reset Threshold |
|---|---|---|---|---|---|
| {TODAY} | Comprehensive | Full codebase | All | Session #{N} - [report link] | ✅ (all) |
2. Update Threshold Summary Table
In the "Current Thresholds" section, update all 6 categories:
- Set "Last Audit" to today's date with "(Comprehensive)"
- Reset "Commits Since" to 0
- Reset "Files Since" to 0
3. Update Multi-AI Thresholds
In the "Multi-AI Audit Thresholds" section:
- Update "Total commits" reset date to today
- Update "Time elapsed" to "0 days (comprehensive audit {TODAY})"
This step ensures
correctly shows no triggers after the
audit.npm run review:check
Triage & Roadmap Integration (MANDATORY)
After TDMS intake completes, triage new items into the roadmap:
1. Review New Items
Check the newly added DEBT-XXXX items:
# View recent additions (last 50 items by ID) tail -50 docs/technical-debt/MASTER_DEBT.jsonl | jq -r '[.id, .severity, .category, .title[:60]] | @tsv'
2. Priority Scoring
Beyond S0-S3 severity, consider these factors for prioritization:
| Factor | Weight | Description |
|---|---|---|
| Severity | 40% | S0=100, S1=50, S2=20, S3=5 |
| Cross-domain | 20% | Items flagged by multiple audits get +50% |
| Effort inverse | 20% | E0=4x, E1=2x, E2=1x, E3=0.5x (quick wins first) |
| Dependency | 10% | Blockers for other items get +25% |
| File hotspot | 10% | Files with 3+ findings get +25% |
Priority Score Formula:
score = (severity × 0.4) × (cross_domain_mult × 0.2) × (effort_inv × 0.2) × (dep_mult × 0.1) × (hotspot_mult × 0.1)
3. Track Assignment
New items are auto-assigned based on category + file patterns:
| Category | File Pattern | Track |
|---|---|---|
| security | * | Track-S |
| performance | * | Track-P |
| process | * | Track-D |
| refactoring | * | M2.3-REF |
| documentation | * | M1.5 |
| code-quality | scripts/, .claude/ | Track-E |
| code-quality | .github/ | Track-D |
| code-quality | tests/ | Track-T |
| code-quality | functions/ | M2.2 |
| code-quality | components/, lib/, app/ | M2.1 |
View current assignments:
cat docs/technical-debt/views/unplaced-items.md
4. Update ROADMAP.md
For S0/S1 items that need immediate attention:
## Track-S: Security Technical Debt - [ ] DEBT-0875: Firebase credentials written to disk (S1) **NEW** - [ ] DEBT-0876: Missing App Check validation (S1) **NEW**
For bulk items by track:
- [ ] DEBT-0869 through DEBT-0880: Process automation gaps (S2, bulk)
5. Consistency Check
Verify all references are valid:
node scripts/debt/sync-roadmap-refs.js --check-only
Reports:
- Orphaned refs (in ROADMAP but not in MASTER_DEBT)
- Unplaced items (in MASTER_DEBT but not in ROADMAP)
- Status mismatches (marked done but not RESOLVED)
6. Review Cadence
| Trigger | Action |
|---|---|
| After comprehensive audit | Full triage of all new items |
| After single-domain audit | Triage items in that category only |
| Weekly (if no audits) | Check unplaced-items.md for drift |
| Before sprint planning | Review S0/S1 items for inclusion |
Completion
Display Final Summary:
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 🎯 COMPREHENSIVE AUDIT COMPLETE ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 📊 Results Summary: • 142 raw findings → 97 unique (45 merged) • S0 Critical: 3 • S1 High: 24 • S2 Medium: 42 • S3 Low: 28 🔍 Cross-Domain Insights: • 8 files need comprehensive refactor (4+ audits) • 12 security/performance overlaps • 5 documentation gaps in complex areas 📄 Full Report: docs/audits/comprehensive/COMPREHENSIVE_AUDIT_REPORT.md ⏱️ Total Time: ~45 minutes (vs 150 minutes sequential - saved ~105 minutes!) 🎯 Recommended Next Steps: 1. Review top 20 priority findings 2. Create GitHub issues for S0/S1 3. Plan refactor for hotspot files
Error Handling
If Individual Audit Fails:
- Continue with remaining audits in the same stage
- Mark failed audit in status display (❌)
- Note failure in final report
- Suggest running failed audit individually for debugging
If Aggregator Fails:
- All individual reports still available
- User can manually review 6 separate reports
- Suggest creating GitHub issue for aggregator failure
If All Audits Fail:
- Check baseline environment (tests passing, lint working)
- Check for system issues (disk space, memory)
- Suggest running single audit first to isolate issue
Context Recovery
If context compacts mid-audit, resume from last completed checkpoint:
Determine Current State
echo "=== Checking audit progress ===" ls -la docs/audits/comprehensive/*.md 2>/dev/null | wc -l
Recovery Matrix
| Files Found | State | Resume Action |
|---|---|---|
| 0-3 reports | Stage 1 incomplete | Re-run missing Stage 1 audits |
| 4 reports | Stage 1 complete | Start Stage 2 |
| 5 reports | Stage 2 incomplete | Re-run missing Stage 2 audit |
| 6 reports, no COMPREHENSIVE | Stage 2 complete | Run Stage 3 (aggregator) |
| COMPREHENSIVE exists | Complete | Run post-audit only |
Resume Commands
Stage 1 incomplete: Re-run only missing audits:
# Check which are missing for audit in code security performance refactoring; do [ ! -f "docs/audits/comprehensive/audit-${audit}-report.md" ] && echo "Missing: $audit" done
Stage 2 incomplete: Run documentation and/or process audits as needed.
Stage 3: Run aggregator on existing 6 reports.
Usage Examples
Quarterly Health Check:
/audit-comprehensive
Pre-Release Audit:
/audit-comprehensive
After Major Refactor:
/audit-comprehensive
Focused Audit (Not Comprehensive):
Use individual skills instead:
/audit-security (25 min - when you only need security review) /audit-code (30 min - when you only need code quality)
Notes
- Staged Execution: 3 stages (4+2+1 agents) respects CLAUDE.md max 4 concurrent limit
- Time Estimate: ~45 minutes (vs 150min sequential = 70% savings)
- S0/S1 Escalation: Security findings checked after Stage 1 before proceeding
- Checkpoints: Each stage verifies outputs exist and are non-empty
- Context Recovery: Can resume from any checkpoint after context compaction
- Output Consistency: All audits use same severity (S0-S3) and effort (E0-E3) scales
Future Enhancements
- Incremental Audits: Only re-run audits for changed domains
- Custom Audit Subset:
to run subset--audits code,security - Confidence Scoring: Weight findings by audit confidence levels
- Trend Analysis: Compare against previous comprehensive audits
- Auto-Issue Creation: Create GitHub issues for S0/S1 findings automatically
Related Skills
- Individual code quality audit/audit-code
- Individual security audit/audit-security
- Individual performance audit/audit-performance
- Individual documentation audit/audit-documentation
- Individual refactoring audit/audit-refactoring
- Individual process/automation audit/audit-process
- Standalone aggregation (if you have existing reports)/audit-aggregator
Documentation References
Before running this audit, review:
TDMS Integration (Required)
- PROCEDURE.md - Full TDMS workflow
- MASTER_DEBT.jsonl - Canonical debt store
- All individual audits automatically run TDMS intake after completion
Documentation Standards (Required)
- JSONL_SCHEMA_STANDARD.md - Output format requirements and TDMS field mapping (used by aggregator)
- DOCUMENTATION_STANDARDS.md - 5-tier doc hierarchy
Version History
| Version | Date | Description |
|---|---|---|
| 2.1 | 2026-02-03 | Added Triage & Roadmap Integration section with priority scoring formula |
| 2.0 | 2026-02-02 | Staged execution (4+2+1), S0/S1 escalation, checkpoints, context recovery |
| 1.0 | 2026-01-28 | Initial version - flat parallel execution of all 6 audits |