Awesome-Agent-Skills-for-Empirical-Research deep-audit
Deep consistency audit of the entire repository — launches 4 parallel specialist agents to find factual errors, code bugs, broken references, count mismatches, and cross-document inconsistencies, then fixes all issues and loops until clean. Make sure to use this skill whenever the user wants a comprehensive repository-wide check — not a targeted review of a single file. Triggers include: "audit", "deep audit", "find inconsistencies", "check everything", "run a full audit", "are there any broken references", "check the whole repo", "something feels off", "run the audit loop", or after making broad changes across multiple files.
git clone https://github.com/brycewang-stanford/Awesome-Agent-Skills-for-Empirical-Research
T=$(mktemp -d) && git clone --depth=1 https://github.com/brycewang-stanford/Awesome-Agent-Skills-for-Empirical-Research "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/15-Felpix-Studios-social-science-research/skills/deep-audit" ~/.claude/skills/brycewang-stanford-awesome-agent-skills-for-empirical-research-deep-audit-724a87 && rm -rf "$T"
skills/15-Felpix-Studios-social-science-research/skills/deep-audit/SKILL.md/deep-audit — Repository Infrastructure Audit
Run a comprehensive consistency audit across the entire repository, fix all issues found, and loop until clean.
When to Use
- After broad changes (new skills, rules, hooks, guide edits)
- Before releases or major commits
- When the user asks to "find inconsistencies", "audit", or "check everything"
Workflow
PHASE 1: Launch 4 Parallel Audit Agents
Launch these 4 agents simultaneously using
Task with subagent_type=general-purpose:
Agent 1: Guide Content Accuracy
Focus:
README.md, CLAUDE.md, rules/workflow-overview.md
- All numeric claims match reality (skill count, agent count, rule count, hook count)
- All file paths mentioned actually exist on disk
- All skill/agent/rule names match actual directory names
- No stale counts from previous versions
Agent 2: Hook Code Quality
Focus:
hooks/*.py and hooks/*.sh
- No remaining
usage (should use/tmp/
)~/.claude/sessions/ - Hash length consistency (
across all hooks)[:8] - Proper error handling (fail-open pattern: top-level
withtry/except
)sys.exit(0) - JSON input/output correctness (stdin for input, stdout/stderr for output)
- Exit code correctness (0 for non-blocking, non-zero only when intentionally blocking)
for Python 3.8+ compatibilityfrom __future__ import annotations- Correct field names from hook input schema (
notsource
for SessionStart)type - PreCompact hooks print to stderr (stdout is ignored)
Agent 3: Skills and Rules Consistency
Focus:
skills/*/SKILL.md and rules/*.md
- Valid YAML frontmatter in all files
- No stale
disable-model-invocation: true
values are sensibleallowed-tools- Rule
reference existing directoriespaths: - No contradictions between rules
- CLAUDE.md skills table matches actual skill directories 1:1
- All templates referenced in rules/guide exist in
templates/
Agent 4: Cross-Document Consistency
Focus:
README.md, CLAUDE.md
- All feature counts agree across both documents
- All links point to valid targets
- Directory tree matches actual structure
- No stale counts from previous versions
PHASE 2: Triage Findings
Categorize each finding:
- Genuine bug: Fix immediately
- False alarm: Discard (document WHY it's false for future rounds)
Common false alarms to watch for:
- Quarto callout
inside## Title
divs — this is standard syntax, NOT a heading bug:::
linter warning — known linter bug (Claude Code issue #25380), field IS validallowed-tools- Counts in old session logs — these are historical records, not user-facing docs
PHASE 3: Fix All Issues
Apply fixes in parallel where possible. For each fix:
- Read the file first (required by Edit tool)
- Apply the fix
- Verify the fix (grep for stale values, check syntax)
PHASE 4: Documentation Check
This plugin does not maintain a Quarto source file. If
README.md or CLAUDE.md were
modified, verify they are consistent with each other — no render step is needed.
PHASE 5: Loop or Declare Clean
After fixing, launch a fresh set of 4 agents to verify.
- If new issues found → fix and loop again
- If zero genuine issues → declare clean and report summary
Max loops: 5 (to prevent infinite cycling)
Key Lessons from Past Audits
These are real bugs found across 7 rounds — check for these specifically:
| Bug Pattern | Where to Check | What Went Wrong |
|---|---|---|
| Stale counts ("19 skills" → "21") | Guide, README, landing page | Added skills but didn't update all mentions |
| Hook exit codes | All Python hooks | Exit 2 in PreCompact silently discards stdout |
| Hook field names | post-compact-restore.py | SessionStart uses , not |
| State in /tmp/ | All Python hooks | Should use |
| Hash length mismatch | All Python hooks | Some used , others |
| Missing fail-open | Python hooks | Unhandled exception → exit 1 → confusing behavior |
| Python 3.10+ syntax | Type hints like `dict | None` |
| Missing directories | quality_reports/specs/ | Referenced in rules but never created |
| Always-on rule listing | Guide + README | meta-governance omitted from listings |
| macOS-only commands | Skills, rules | without fallback |
| Protected file blocking | settings.json edits | protect-files.sh blocks Edit/Write |
Output Format
After each round, report:
## Round N Audit Results ### Issues Found: X genuine, Y false alarms | # | Severity | File | Issue | Status | |---|----------|------|-------|--------| | 1 | Critical | file.py:42 | Description | Fixed | | 2 | Medium | file.qmd:100 | Description | Fixed | ### Verification - [ ] No stale counts (grep confirms) - [ ] All hooks have fail-open + future annotations - [ ] Guide renders successfully - [ ] docs/ updated ### Result: [CLEAN | N issues remaining]