Awesome-Agent-Skills-for-Empirical-Research parsifal-slr-guide
Plan and manage systematic literature reviews with Parsifal platform
install
source · Clone the upstream repo
git clone https://github.com/brycewang-stanford/Awesome-Agent-Skills-for-Empirical-Research
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/brycewang-stanford/Awesome-Agent-Skills-for-Empirical-Research "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/43-wentorai-research-plugins/skills/research/methodology/parsifal-slr-guide" ~/.claude/skills/brycewang-stanford-awesome-agent-skills-for-empirical-research-parsifal-slr-guid && rm -rf "$T"
manifest:
skills/43-wentorai-research-plugins/skills/research/methodology/parsifal-slr-guide/SKILL.mdsource content
Parsifal Systematic Literature Review Guide
Overview
Parsifal is a web-based tool for planning and managing systematic literature reviews (SLRs) following established protocols (Kitchenham, PRISMA). It guides researchers through the complete SLR process: defining research questions, setting inclusion/exclusion criteria, planning search strings, and tracking the screening process. Open-source and self-hostable.
SLR Process with Parsifal
Phase 1: Planning
Define Research Questions
Structure questions using PICO framework:
- Population: What group/domain?
- Intervention: What technique/method?
- Comparison: Compared to what?
- Outcome: What results measured?
Example: P: Software development teams I: AI-assisted code review C: Manual code review O: Defect detection rate, review time Research Questions: RQ1: Does AI-assisted code review improve defect detection? RQ2: What is the time savings compared to manual review? RQ3: What types of defects are best detected by AI tools?
Set Criteria
Inclusion Criteria: IC1: Studies comparing AI vs manual code review IC2: Published in peer-reviewed venues (2020-2026) IC3: Reports quantitative metrics Exclusion Criteria: EC1: Grey literature / blog posts EC2: Studies with fewer than 10 participants EC3: Non-English publications
Phase 2: Search Strategy
Build Search String
("artificial intelligence" OR "machine learning" OR "deep learning") AND ("code review" OR "code inspection" OR "static analysis") AND ("defect detection" OR "bug finding" OR "software quality")
Database Mapping
| Database | Adapted Query | Expected Results |
|---|---|---|
| Scopus | TITLE-ABS-KEY(...) | ~500 |
| IEEE Xplore | querytext=... | ~300 |
| ACM DL | [[Abstract: ...]] | ~200 |
| Web of Science | TS=(...) | ~400 |
Phase 3: Selection
Screening Steps
- Remove duplicates — Match by DOI, title similarity
- Title screening — Quick relevance assessment
- Abstract screening — Apply inclusion/exclusion criteria
- Full-text review — Detailed evaluation
Quality Assessment
Define quality criteria and scoring:
| Criterion | Score |
|---|---|
| Clear research question stated | 0/0.5/1 |
| Methodology described in detail | 0/0.5/1 |
| Threats to validity discussed | 0/0.5/1 |
| Results statistically analyzed | 0/0.5/1 |
| Study replicable from description | 0/0.5/1 |
Phase 4: Extraction
Data Extraction Form
For each included paper, extract: - Study ID - Authors, Year, Venue - Study type (experiment/case study/survey) - Population size - AI technique used - Metrics reported (precision, recall, F1, time) - Key findings - Limitations noted
Phase 5: Synthesis
Report with PRISMA
Identification: 1,400 records ↓ Remove duplicates: -350 Screening: 1,050 titles/abstracts ↓ Exclude irrelevant: -900 Eligibility: 150 full-text assessed ↓ Exclude by criteria: -108 Included: 42 studies in final review
Self-Hosting Parsifal
git clone https://github.com/vitorfs/parsifal.git cd parsifal pip install -r requirements.txt python manage.py migrate python manage.py runserver # Access at http://localhost:8000
SLR Best Practices
- Register protocol before starting (PROSPERO for health, OSF for others)
- Two independent reviewers for screening to reduce bias
- Track inter-rater agreement (Cohen's kappa > 0.8)
- Document deviations from the original protocol
- Use PRISMA checklist for reporting completeness
References
- Parsifal
- Kitchenham, B. & Charters, S. (2007). "Guidelines for performing Systematic Literature Reviews in Software Engineering."
- PRISMA Statement
- PROSPERO Registry