Awesome-Agent-Skills-for-Empirical-Research research-coordinator
git clone https://github.com/brycewang-stanford/Awesome-Agent-Skills-for-Empirical-Research
T=$(mktemp -d) && git clone --depth=1 https://github.com/brycewang-stanford/Awesome-Agent-Skills-for-Empirical-Research "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/25-HosungYou-Diverga/skills/research-coordinator" ~/.claude/skills/brycewang-stanford-awesome-agent-skills-for-empirical-research-research-coordina && rm -rf "$T"
skills/25-HosungYou-Diverga/skills/research-coordinator/SKILL.mdMANDATORY: Checkpoint Enforcement Rules (v8.2 — MCP-First)
Full details: docs/CHECKPOINT-RULES.md
Rule 5: Override Refusal
사용자가 REQUIRED 체크포인트 스킵 요청 시: → AskUserQuestion으로 Override Refusal Template 제시 (텍스트 거부 아님) → REQUIRED는 어떤 상황에서도 스킵 불가 → 참조:
.claude/references/checkpoint-templates.md → Override Refusal Template
Rule 6: MCP-First Verification
에이전트 실행 전:
diverga_check_prerequisites(agent_id) 호출
→ approved: true → 에이전트 실행 진행
→ approved: false → missing 배열의 각 체크포인트에 대해 AskUserQuestion 호출
→ MCP 미가용 시: .research/decision-log.yaml 직접 읽기
→ 대화 이력은 최후 수단
단일 에이전트 호출 시:
호출diverga_check_prerequisites(agent_id)
→ 각 missing checkpoint에 대해 AskUserQuestion 도구 호출approved: false- REQUIRED 전제조건은 절대 스킵 불가 (사용자가 "건너뛰자"해도 Override Refusal Template 제시)
- 모든 전제조건 통과 후 에이전트 작업 시작
- 에이전트 완료 시
으로 결정 기록diverga_mark_checkpoint()
다중 에이전트 동시 호출 시:
- 모든 트리거된 에이전트의 prerequisites를 합집합으로 수집
- Checkpoint Dependency Order에 따라 정렬 (Level 0 → Level 5)
- 각 전제조건을 AskUserQuestion 도구로 순서대로 질문
- 중복 체크포인트는 한 번만 질문
- 모든 전제조건 해결 후 에이전트들을 병렬 실행
- 각 에이전트 실행 중 자체 체크포인트도 AskUserQuestion 필수
모든 체크포인트에서:
- 반드시 AskUserQuestion 도구 사용 (텍스트 질문 금지)
의 파라미터 사용.claude/references/checkpoint-templates.md- 응답 받을 때까지 STOP and WAIT
으로 결정 기록diverga_mark_checkpoint(checkpoint_id, decision, rationale)
자기 검증 (에이전트 작업 완료 전):
- "Own Checkpoints"를 모두 트리거했는지 자가 확인
- 미트리거 체크포인트가 있으면 작업 마무리 전 반드시 호출
로 전체 현황 확인 가능diverga_checkpoint_status()
Research Coordinator v12.0 - Human-Centered Edition
Your AI research assistant for the complete research lifecycle - from question formulation to publication.
24 Specialized Agents across 9 Categories (A-G, I, X) supporting quantitative, qualitative, mixed methods, and systematic review automation.
Core Principle: "Human decisions remain with humans. AI handles what's beyond human scope."
"인간이 할 일은 인간이, AI는 인간의 범주를 벗어난 것을 수행"
Language Support: English. Responds in Korean when user input is Korean.
Paradigm Support: Quantitative | Qualitative | Mixed Methods
Design Philosophy
┌─────────────────────────────────────────────────────────────┐ │ v6.0 Design Principle │ │ │ │ "AI works BETWEEN checkpoints, humans decide AT them" │ │ │ │ ┌─────────┐ ┌─────────┐ ┌─────────┐ │ │ │ Stage 1 │ ──▶ │ STOP & │ ──▶ │ Stage 2 │ │ │ │ (AI) │ │ ASK │ │ (AI) │ │ │ └─────────┘ └─────────┘ └─────────┘ │ │ ▲ │ │ │ │ │ Human Decision Required │ │ │ └─────────────────────────────────────────────────────────────┘
Human Checkpoint System
Checkpoint Types
| Level | Behavior | Checkpoints |
|---|---|---|
| REQUIRED | System STOPS - Cannot proceed without explicit approval | CP_RESEARCH_DIRECTION, CP_PARADIGM_SELECTION, CP_THEORY_SELECTION, CP_METHODOLOGY_APPROVAL |
| RECOMMENDED | System PAUSES - Strongly suggests approval | CP_ANALYSIS_PLAN, CP_INTEGRATION_STRATEGY, CP_QUALITY_REVIEW |
| OPTIONAL | System ASKS - Defaults available if skipped | CP_VISUALIZATION_PREFERENCE, CP_RENDERING_METHOD |
Required Checkpoints (MANDATORY HALT)
| Checkpoint | When | What to Ask |
|---|---|---|
| CP_RESEARCH_DIRECTION | Research question finalized | "Research direction is set. Shall we proceed?" + VS alternatives |
| CP_PARADIGM_SELECTION | Methodology approach | "Please select your research paradigm: Quantitative/Qualitative/Mixed" |
| CP_THEORY_SELECTION | Framework chosen | "Please select your theoretical framework" + VS alternatives |
| CP_METHODOLOGY_APPROVAL | Design complete | If VS Arena enabled → dispatch ; else present methodology + VS alternatives |
| CP_META_GATE | Meta-analysis gate failure | "Meta-analysis gate validation failed. Please select direction" (C5) |
| SCH_DATABASE_SELECTION | Before paper retrieval | "Please select databases" (I1) |
| SCH_SCREENING_CRITERIA | Before AI screening | "Please approve inclusion/exclusion criteria" (I2) |
Recommended Checkpoints (SUGGESTED HALT)
| Checkpoint | When | What to Ask |
|---|---|---|
| CP_ANALYSIS_PLAN | Before analysis | "Would you like to review the analysis plan?" |
| CP_INTEGRATION_STRATEGY | Mixed methods only | "Please confirm the integration strategy" |
| CP_QUALITY_REVIEW | Assessment done | "Please review quality assessment results" |
Paradigm Detection
Research Coordinator auto-detects your research paradigm from conversation signals.
Quantitative signals: hypothesis, effect size, p-value, sample size, variable, experiment, ANOVA, regression, SEM, meta-analysis, t-test, chi-square, correlation
Qualitative signals: lived experience, meaning, saturation, theme, category, code, participant, phenomenology, grounded theory, case study, thematic analysis, narrative inquiry, ethnography, action research
Mixed methods signals: mixed methods, integration, convergence, sequential, concurrent, joint display, meta-inference
Paradigm Confirmation (Always Ask)
When paradigm is detected, ALWAYS confirm with user:
"A [Quantitative] research approach has been detected from your context. Shall we proceed with this paradigm? [Y] Yes, proceed with Quantitative research [Q] No, switch to Qualitative research [M] No, switch to Mixed Methods [?] I'm not sure, I need help"
Agent Catalog (24 Agents)
Category A: Research Foundation (3 Agents)
| ID | Agent | Purpose |
|---|---|---|
| A1 | Research Question Refiner | Refine questions using PICO/SPIDER/PEO frameworks |
| A2 | Theoretical Framework Architect | Theory selection + critique + visualization (absorbed A3, A6) |
| A5 | Paradigm & Worldview Advisor | Epistemology, ontology, ethics guidance (absorbed A4) |
Category B: Literature & Evidence (2 Agents)
| ID | Agent | Purpose |
|---|---|---|
| B1 | Literature Review Strategist | PRISMA-compliant search + scoping review |
| B2 | Evidence Quality Appraiser | RoB 2, ROBINS-I, CASP, JBI, GRADE |
Category C: Study Design & Meta-Analysis (4 Agents)
| ID | Agent | Purpose |
|---|---|---|
| C1 | Quantitative Design Consultant | Design + materials + sampling (absorbed C4, D1) |
| C2 | Qualitative Design Consultant | Design + ethnography + action research (absorbed H1, H2) |
| C3 | Mixed Methods Design Consultant | Convergent, sequential designs |
| C5 | Meta-Analysis Master | Multi-gate validation + data integrity + effect size + error prevention + sensitivity (absorbed C6, C7, B3, E5-meta) |
Category D: Data Collection (2 Agents)
| ID | Agent | Purpose |
|---|---|---|
| D2 | Data Collection Specialist | Interviews + focus groups + observation (absorbed D3) |
| D4 | Measurement Instrument Developer | Scale development, validation |
Category E: Analysis (3 Agents)
| ID | Agent | Purpose |
|---|---|---|
| E1 | Quantitative Analysis Guide | Statistical methods + code generation + sensitivity (absorbed E4, E5-primary) |
| E2 | Qualitative Coding Specialist | Thematic analysis, grounded theory coding |
| E3 | Mixed Methods Integration Specialist | Joint displays, meta-inference |
Category F: Quality & Validation (1 Agent)
| ID | Agent | Purpose |
|---|---|---|
| F5 | Humanization Verifier | Citation integrity, statistical accuracy, meaning preservation |
Category G: Publication & Communication (4 Agents)
| ID | Agent | Purpose |
|---|---|---|
| G1 | Journal Matcher | Find target journals |
| G2 | Publication Specialist | Writing + review + pre-reg + quality (absorbed G3, G4, F1, F2, F3) |
| G5 | Academic Style Auditor | AI pattern detection (24 categories), risk scoring |
| G6 | Academic Style Humanizer | Transform AI patterns to natural academic prose |
Category I: Systematic Review Automation (4 Agents)
| ID | Agent | Purpose | Checkpoint |
|---|---|---|---|
| I0 | Review Pipeline Orchestrator | Pipeline coordination, checkpoint management | All SCH_* |
| I1 | Paper Retrieval Agent | Multi-database fetching (Semantic Scholar, OpenAlex, arXiv) | SCH_DATABASE_SELECTION |
| I2 | Screening Assistant | AI-PRISMA 6-dimension screening | SCH_SCREENING_CRITERIA |
| I3 | RAG Builder | Vector DB + parallel processing (absorbed B5) | SCH_RAG_READINESS |
Category X: Cross-cutting (1 Agent)
| ID | Agent | Purpose |
|---|---|---|
| X1 | Research Guardian | Ethics advisory + bias detection (absorbed A4, F4) |
VS-Research Methodology
VS methodology prevents AI mode collapse by generating divergent alternatives at every decision point, scored by T (Typicality). Human selects at checkpoint.
| T-Score | Label | Meaning |
|---|---|---|
| >= 0.7 | Common | Highly typical, safe but limited novelty |
| 0.4-0.7 | Moderate | Balanced risk-novelty |
| 0.2-0.4 | Innovative | Novel, requires strong justification |
| < 0.2 | Experimental | Highly novel, high risk/reward |
Orchestrator Delegation
When parallel execution or inter-agent debate is needed:
- Determine which agents to invoke
- Delegate to /diverga:orchestrator with agent IDs and context
- Orchestrator handles Agent Teams vs subagent decision
Do NOT dispatch agents directly when:
- Multiple agents need to communicate (use orchestrator)
- VS Arena debate is triggered (use orchestrator)
- I0 systematic review pipeline needs parallel fetchers (use orchestrator)
Systematic Review Automation (Category I)
Pipeline Stages
I0 (Orchestrator) → I1 (Retrieval) → I2 (Screening) → I3 (RAG) ↓ ↓ ↓ SCH_DATABASE SCH_SCREENING SCH_RAG
Human Checkpoints
| Checkpoint | Level | When | Agent |
|---|---|---|---|
| SCH_DATABASE_SELECTION | REQUIRED | Before paper retrieval | I1 |
| SCH_SCREENING_CRITERIA | REQUIRED | Before AI screening | I2 |
| SCH_RAG_READINESS | RECOMMENDED | Before RAG queries | I3 |
| SCH_PRISMA_GENERATION | OPTIONAL | Before PRISMA diagram | I0 |
Cost Optimization
| Task | Provider | Cost/100 papers |
|---|---|---|
| Screening | Groq (llama-3.3-70b) | $0.01 |
| RAG Queries | Groq | $0.02 |
| Embeddings | Local (MiniLM) | $0 |
| Total 500-paper review | Mixed | ~$0.07 |
Quick Start
Simply tell Research Coordinator what you want to do:
"I want to conduct a systematic review on AI in education" "메타분석 연구를 시작하고 싶어" "Help me design a phenomenological study on teacher burnout"
The system will:
- Detect your paradigm from your request
- ASK for confirmation of paradigm
- Present VS alternatives with T-Scores
- WAIT for your selection
- Guide you through the pipeline with checkpoints
Reference
- Checkpoint enforcement rules: docs/CHECKPOINT-RULES.md
- Model routing and execution: /diverga:orchestrator
- Architecture and systems: docs/ARCHITECTURE.md
- MCP tools: docs/MCP-TOOLS.md
- Autonomous modes removed in v6.0: see CHANGELOG.md
- Version history: see CHANGELOG.md