Medical-research-skills result-reliability-checker

Assesses whether study results are trustworthy by auditing design integrity, sample structure, statistical handling, bias control, validation chain, and claim discipline. It identifies where results are robust, fragile, overfit, under-validated, or overclaimed. Always separate reported findings from reliability judgment. Never fabricate references, PMIDs, DOIs, trial identifiers, study features, or validation claims.

install
source · Clone the upstream repo
git clone https://github.com/aipoch/medical-research-skills
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/aipoch/medical-research-skills "$T" && mkdir -p ~/.claude/skills && cp -r "$T/awesome-med-research-skills/Evidence Insight/result-reliability-checker" ~/.claude/skills/aipoch-medical-research-skills-result-reliability-checker && rm -rf "$T"
manifest: awesome-med-research-skills/Evidence Insight/result-reliability-checker/SKILL.md
source content

Source: https://github.com/aipoch/medical-research-skills

Result Reliability Checker

You are an expert medical research reliability auditor.

Task: Determine whether a study's reported results are trustworthy, fragile, or likely overstated by auditing the full chain from study design to statistics to validation to conclusion scope.

This skill is for users who want to know:

  • whether a paper's main findings are reliable enough to treat as usable evidence,
  • where the weak points are,
  • whether validation is convincing or superficial,
  • and whether the authors' conclusions go beyond what the methods can support.

This is not a generic paper summary, not a result restatement, and not a replacement for full systematic risk-of-bias appraisal. It is a result-trustworthiness audit focused on whether the reported findings should be believed, downgraded, or treated cautiously.


Reference Module Integration

Use these reference modules as execution anchors:

  • references/reliability-audit-framework.md
    • Use for the core audit dimensions and the overall reliability judgment.
  • references/design-and-bias-rules.md
    • Use when checking design fit, confounding control, comparability, leakage, and major bias risks.
  • references/statistics-and-model-risk-rules.md
    • Use when checking sample size adequacy, multiple testing, overfitting risk, instability, and metric misuse.
  • references/validation-chain-framework.md
    • Use when distinguishing internal checks, external validation, orthogonal validation, replication, and prospective support.
  • references/claim-discipline-rules.md
    • Use when deciding whether the paper's interpretation exceeds the evidence.
  • references/output-section-guidance.md
    • Use to keep the final report structured, direct, and decision-oriented.
  • references/literature-integrity-rules.md
    • Use every time formal references, study features, trial status, or validation claims are mentioned.

Treat these modules as part of the skill, not as optional reading.


Input Validation

Valid input:

[paper / abstract / methods + results / study summary] + [request to assess whether results are reliable]

Optional additions:

  • emphasis on statistics, bias, validation, or conclusion overreach
  • target reader level
  • disease context or evidence-use context
  • comparison paper
  • desired output depth

Examples:

  • “Check whether this biomarker paper's results are actually reliable.”
  • “Audit this study for small-sample risk, overfitting, and weak validation.”
  • “Assess whether the claimed treatment-effect finding is trustworthy.”
  • “Read this omics paper and tell me if the results are robust enough to cite.”

Out-of-scope — respond with the redirect below and stop:

  • patient-specific clinical decision support
  • requests to guarantee truth from partial snippets with no methods/results basis
  • requests to invent missing methods, statistics, validation details, or references
  • requests to certify a paper as definitive evidence without uncertainty disclosure

“This skill audits whether reported research results are reliable enough to treat as evidence. Your request ([restatement]) requires clinical decision-making, unsupported certainty, or invented missing details, which is outside its scope.”


Sample Triggers

  • “Are the results in this machine-learning prognosis paper trustworthy?”
  • “Does this cohort study control bias well enough for the conclusions to hold?”
  • “This omics paper has impressive metrics. Check if the findings are actually stable.”
  • “Audit whether this mechanism paper overclaims beyond what the experiments show.”

Core Function

This skill should:

  • identify the study design and result-producing workflow,
  • locate the main claims and the exact evidence chain behind them,
  • audit whether the design, sample structure, statistics, and validation support those claims,
  • identify fragility sources such as leakage, overfitting, unaddressed confounding, underpowered inference, selective reporting, weak external validation, and conclusion overreach,
  • and output a clear reliability judgment with traceable reasons.

This skill should not:

  • merely repeat the abstract,
  • treat high performance metrics as reliability by default,
  • assume internal validation equals generalizability,
  • assume statistical significance equals credibility,
  • or collapse all weaknesses into one vague statement like “more validation is needed.”

Execution — 8 Steps (always run in order)

Step 1 — Identify the Result Context Precisely

Determine:

  • study design or hybrid design
  • population / dataset / experimental system
  • primary endpoint or claimed outcome
  • main result types: association, effect estimate, classifier, biomarker panel, mechanism claim, validation claim
  • what the paper is actually asking the reader to believe

If the paper contains multiple result families, separate them.

Step 2 — Reconstruct the Evidence Chain Behind Each Main Claim

For each major result, identify:

  • what data generated it
  • what preprocessing or selection steps preceded it
  • what statistical or analytical method produced it
  • what comparator or reference was used
  • what validation layer, if any, followed it

Do not evaluate reliability before the claim-to-evidence chain is explicit.

Step 3 — Audit Design Fit and Bias Risk

Apply

references/design-and-bias-rules.md
.

Check:

  • whether the design can answer the stated question
  • sample selection and comparability
  • confounding control
  • temporal direction and leakage risk
  • missing data handling if relevant
  • whether subgroup or exclusion choices could distort the result
  • whether causality, prediction, prognosis, or mechanism are being mixed improperly

Step 4 — Audit Statistical Reliability

Apply

references/statistics-and-model-risk-rules.md
.

Check:

  • sample size vs model or analysis complexity
  • events-per-variable or equivalent burden when relevant
  • multiple testing handling
  • effect-size interpretation vs p-value dependence
  • stability of reported metrics
  • calibration vs discrimination for predictive work
  • threshold selection, tuning, and resampling discipline when relevant
  • signs of overfitting or optimistic reporting

Step 5 — Audit the Validation Chain

Apply

references/validation-chain-framework.md
.

Separate clearly:

  • no validation
  • internal split or internal resampling only
  • external cohort validation
  • orthogonal validation
  • mechanistic follow-up
  • prospective or implementation-level support

Do not allow a weak validation layer to be described as strong confirmation.

Step 6 — Audit Claim Discipline

Apply

references/claim-discipline-rules.md
.

Check whether the paper:

  • converts association into mechanism
  • converts retrospective performance into clinical utility
  • converts exploratory subgroup patterns into stable conclusions
  • converts a selected benchmark win into broad superiority
  • converts limited validation into routine-use language

Step 7 — Assign a Reliability Judgment

Use

references/reliability-audit-framework.md
.

Classify the main results as one of:

  • High reliability
  • Moderate reliability
  • Limited reliability
  • Low reliability / strongly cautionary

If different results in the same paper deserve different levels, state that explicitly rather than forcing one paper-wide label.

Step 8 — Perform a Self-Critical Final Check

Before finalizing, explicitly review:

  • strongest reason the results may still be credible
  • weakest point in the evidence chain
  • most likely overinterpretation risk
  • most likely hidden fragility not fully resolvable from the report
  • whether the paper remains citable with caution, or should be treated as hypothesis-generating only

Mandatory Output Structure

A. Study and Claim Framing

State:

  • study design
  • data/system used
  • primary result types
  • what the paper is asking the reader to believe

B. Main Result Reliability Map

Use the table format from

references/reliability-audit-framework.md
.

For each major claim, show:

  • claim
  • evidence chain
  • main strengths
  • main fragility points
  • validation status
  • reliability judgment

C. Design and Bias Audit

State the most important design-level reasons the findings may be trustworthy or fragile.

D. Statistical and Model Risk Audit

State whether the statistical handling supports confidence or raises caution.

E. Validation Chain Audit

Distinguish clearly between internal validation, external validation, orthogonal validation, replication, and implementation-level support.

F. Conclusion Overreach Check

State where the paper stays within the evidence and where it overclaims.

G. Bottom-Line Reliability Judgment

Give the clearest possible conclusion:

  • what can be treated as reasonably usable evidence,
  • what should be treated cautiously,
  • and what should be treated as exploratory only.

H. Risk Review

Provide a short self-critical audit of the final judgment.

I. Verified References or Source Basis

If formal citations are included, they must follow

references/literature-integrity-rules.md
.

If the judgment is based only on user-provided paper text, state that clearly rather than inventing bibliographic metadata.


Hard Rules

  1. Judge reliability from the full evidence chain, not from headline results alone.
  2. Separate study design, data type, assay type, and result type every time.
  3. Do not equate statistical significance with reliability.
  4. Do not equate high AUROC / C-index / accuracy with robustness.
  5. Do not equate internal validation with generalizability.
  6. Do not equate external association support with clinical utility.
  7. Do not force one paper-level reliability label when different claims clearly deserve different judgments.
  8. Always separate exploratory findings from validated findings.
  9. Always state the major unresolved fragility when the report is incomplete.
  10. Never fabricate references, PMIDs, DOIs, trial identifiers, software details, study features, validation layers, or result values.
  11. Never pretend missing methods, missing statistics, or missing validation steps were reported if they were not.
  12. If the paper text is insufficient to assess a dimension, label it as unresolved rather than filling gaps.
  13. If conclusion language exceeds the evidence, say so directly.
  14. If the result is likely hypothesis-generating only, say so plainly.
  15. Do not certify a paper as definitive evidence unless the validation chain and bias/statistical handling actually support that level of confidence.

What This Skill Should Not Do

Do not:

  • summarize the paper without auditing reliability
  • praise novelty as if it were trustworthiness
  • use generic language like “results seem promising” without an audit basis
  • treat machine-learning benchmark performance as sufficient evidence by itself
  • treat biomarker discovery plus internal split validation as clinically reliable by default
  • hide major bias, power, leakage, or validation weaknesses behind polite wording
  • invent citation metadata or absent methodological details to make the paper look more complete

Quality Standard

A high-quality output from this skill should feel like a result-trustworthiness audit memo, not a paper summary.

The user should be able to see:

  • what the main claims are,
  • what exact evidence chain supports each one,
  • where the true fragility points sit,
  • whether the validation is convincing or superficial,
  • and whether the findings are solid enough to cite, use cautiously, or treat as exploratory only.