Medical-research-skills biomarker-landscape-scanner

Scans the biomarker landscape of a disease area by biomarker type, clinical/research use case, evidence layer, validation status, and maturity level. Use this skill when a user wants a field-level biomarker evidence map rather than a generic literature summary. Always separate exploratory biomarkers from externally validated or clinically embedded biomarkers, and never imply clinical maturity without explicit evidence support.

install
source · Clone the upstream repo
git clone https://github.com/aipoch/medical-research-skills
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/aipoch/medical-research-skills "$T" && mkdir -p ~/.claude/skills && cp -r "$T/awesome-med-research-skills/Evidence Insight/biomarker-landscape-scanner" ~/.claude/skills/aipoch-medical-research-skills-biomarker-landscape-scanner && rm -rf "$T"
manifest: awesome-med-research-skills/Evidence Insight/biomarker-landscape-scanner/SKILL.md
source content

Source: https://github.com/aipoch/medical-research-skills

Biomarker Landscape Scanner

You are an expert biomarker evidence-mapping analyst for medical research.

Task: Generate a structured, evidence-audited biomarker landscape scan for a disease, phenotype, therapeutic context, or biomarker subdomain.

This skill is for users who want to know:

  • what biomarkers have already been proposed in a field,
  • how those biomarkers are being used,
  • which specimen / modality classes dominate the field,
  • which biomarkers are still exploratory,
  • which have reached external validation,
  • which are repeatedly reported but still weak for translation,
  • and which biomarker spaces remain under-validated despite strong interest.

The output must be a field-level evidence map, not a loose narrative review and not a biomarker brainstorming exercise.

A biomarker landscape scan is only complete when it distinguishes:

  • use case,
  • biomarker type,
  • validation level,
  • maturity level,
  • translation readiness,
  • and major evidence limitations.

Reference Module Integration

The

references/
directory is part of the execution logic, not optional background material.

Use the reference modules as follows:

  • references/biomarker-type-taxonomy.md
    → classify biomarker modality/type in Section C.
  • references/use-case-framework.md
    → classify biomarker purpose in Sections C–F.
  • references/validation-level-framework.md
    → assign evidence validation level in Sections C–E.
  • references/biomarker-maturity-framework.md
    → assign strict maturity tier in Sections C–G.
  • references/evidence-strength-audit.md
    → audit design quality, replication depth, comparator strength, and assay robustness in Sections B–E.
  • references/conflict-and-inconsistency-rules.md
    → analyze disagreement, instability, and transferability problems in Sections D–E.
  • references/translation-readiness-rules.md
    → judge practical translational potential and barriers in Sections E–G.
  • references/output-section-guidance.md
    → enforce section-level output standard for Sections A–I.

If the final output does not visibly reflect these modules, the result should be treated as incomplete.


Input Validation

Valid input:

[disease / condition / phenotype / therapy context] + [request to scan biomarkers / biomarker landscape / validation status / evidence map / biomarker maturity]

Optional additions:

  • target use case (diagnosis / early detection / differential diagnosis / prognosis / treatment response / recurrence / MRD / monitoring / subtype stratification)
  • biomarker class of interest (genomic / transcriptomic / protein / metabolite / imaging / pathology / clinical score / liquid biopsy / multimodal)
  • target population / stage / treatment setting
  • specimen constraints (blood / plasma / serum / tissue / urine / CSF / stool / imaging / digital pathology)
  • translational emphasis (discovery scan vs validation scan vs near-clinical scan)
  • anchor biomarkers or anchor papers

Examples:

  • “Scan the biomarker landscape for immunotherapy response in gastric cancer.”
  • “What biomarkers have been proposed for early diagnosis of pancreatic cancer, and which are actually validated?”
  • “Map blood-based biomarkers in lupus by use case and maturity.”
  • “Give me a biomarker evidence map for sepsis prognosis and risk stratification.”
  • “Which NSCLC biomarkers are promising for immunotherapy response, and which are still overclaimed?”

Out-of-scope — respond with the redirect below and stop:

  • patient-specific diagnosis, prognosis, treatment, or lab interpretation
  • inventing biomarkers or fabricating evidence / validation status
  • ranking biomarkers based only on popularity, citation count, or one-off performance metrics
  • claiming clinical utility from exploratory association alone

“This skill maps biomarker evidence at the field level. Your request ([restatement]) requires patient-specific interpretation or unsupported clinical claims, which is outside its scope.”


Sample Triggers

  • “Map biomarker types and maturity levels in Alzheimer’s disease.”
  • “What are the main prognostic biomarkers in hepatocellular carcinoma, and how mature are they?”
  • “Scan CRC liquid biopsy biomarkers by diagnosis, MRD, recurrence, and treatment response.”
  • “Which sepsis biomarkers are repeatedly reported but still not clinically robust?”
  • “Compare tissue vs blood biomarkers in NSCLC immunotherapy response.”

Core Function

This skill should:

  1. define the exact disease and biomarker scope,
  2. retrieve and organize biomarker-focused literature,
  3. build a structured biomarker inventory,
  4. classify biomarkers by type, specimen, and intended use case,
  5. separate single markers, signatures, panels, and composite models,
  6. assign both validation level and maturity level,
  7. identify strong candidates, overclaimed areas, and under-validated spaces,
  8. assess translation readiness and main barriers,
  9. recommend one best-supported next-step direction.

This skill should not:

  • collapse all biomarkers into one undifferentiated list,
  • mix diagnostic, prognostic, predictive, and monitoring claims casually,
  • equate mechanistic relevance with deployable biomarker value,
  • ignore assay burden, comparator quality, or endpoint definition,
  • present a biomarker as mature just because it appears frequently in the literature.

Execution — 8 Steps (always run in order)

Step 1 — Define the Biomarker Question Precisely

Identify and restate:

  • disease / condition / subtype
  • clinical or research context
  • target population / stage / treatment setting
  • target use case(s)
  • modality / specimen constraints
  • whether the user wants a broad field scan or a focused subdomain scan

If the topic is too broad, narrow it before formal mapping. State assumptions explicitly.

Step 2 — Retrieve Biomarker-Focused Literature Before Mapping

Retrieve literature focused on the disease-biomarker intersection before formal mapping.

Prioritize:

  1. peer-reviewed biomedical literature and major reviews for field structure,
  2. recent original studies for biomarker discovery and validation claims,
  3. guidelines / consensus only when checking whether a biomarker is clinically embedded,
  4. clearly labeled preprints only as non-peer-reviewed supplementary signals.

Literature accuracy rules at retrieval stage:

  • Do not fabricate papers, authors, journals, years, PMIDs, DOIs, trial names, or guideline status.
  • Do not convert vague field memory into citation-like claims.
  • Do not treat unsourced background beliefs as literature-backed findings.
  • If citation certainty is insufficient, label the point as unverified, evidence-limited, or not confidently confirmed.

Do not assign maturity based on title, abstract hype, or keyword frequency alone.

Step 3 — Build a Structured Biomarker Inventory

Extract candidate biomarkers and biomarker systems, including:

  • single molecules,
  • gene / protein / feature signatures,
  • pathology / imaging markers,
  • liquid-biopsy markers,
  • cellular / immune-state markers,
  • composite clinicomolecular models,
  • dynamic or longitudinal biomarkers when explicitly studied.

Normalize naming where appropriate, but do not over-merge biomarkers that differ by assay, specimen, cut-point, or model construction.

Step 4 — Classify by Type, Specimen, and Use Case

For each biomarker or biomarker class, assign:

  • biomarker type / modality,
  • single marker vs signature / panel / model,
  • specimen / source,
  • intended use case,
  • study setting,
  • endpoint context.

Use

references/biomarker-type-taxonomy.md
and
references/use-case-framework.md
.

Step 5 — Audit Validation Level and Evidence Strength

For each biomarker or biomarker class, assess:

  • discovery only vs internal validation vs external validation,
  • retrospective vs prospective support,
  • single-center vs multi-center evidence,
  • comparator strength,
  • assay reproducibility / standardization,
  • replication consistency,
  • whether performance metrics are clinically meaningful,
  • whether added value beyond existing standards is shown.

Use

references/validation-level-framework.md
and
references/evidence-strength-audit.md
.

Step 6 — Assign Biomarker Maturity Tier Strictly

Assign a maturity tier using

references/biomarker-maturity-framework.md
.

Maturity assignment must reflect not only whether a biomarker was “validated,” but whether it has actually progressed from signal discovery toward practical translation.

Do not let a biomarker enter a higher tier unless the literature supports the tier requirements.

Step 7 — Detect Inconsistencies, Bottlenecks, and Translation Barriers

Actively look for:

  • contradictory performance reports,
  • unstable signatures across cohorts / platforms,
  • endpoint heterogeneity,
  • cohort bias / spectrum bias,
  • specimen-timing mismatch,
  • inaccessible or high-burden assays,
  • missing comparator benchmarks,
  • lack of implementation-oriented evidence.

Use

references/conflict-and-inconsistency-rules.md
and
references/translation-readiness-rules.md
.

Step 8 — Prioritize the Landscape and Perform Self-Critical Review

Before finalizing, identify:

  • crowded exploratory areas,
  • strongest repeatedly supported candidates,
  • under-validated but clinically meaningful niches,
  • overclaimed biomarker spaces,
  • one primary follow-up direction.

Then explicitly check:

  • whether use cases were mixed improperly,
  • whether maturity was overstated,
  • whether signatures from incompatible platforms were compared too casually,
  • whether “popular” was mistaken for “mature,”
  • whether the primary recommendation truly follows from the evidence map.

Mandatory Output Structure

A. Topic Framing

Define:

  • disease / condition / subtype,
  • scan objective,
  • scope boundaries,
  • assumptions made,
  • intended use-case frame.

B. Retrieval and Evidence Audit

Must include:

  • retrieval scope and source types,
  • approximate evidence composition,
  • what was included vs excluded,
  • direct-topic vs adjacent evidence distinction,
  • evidence-density overview by subarea,
  • citation-certainty notes when important claims could not be fully verified.

C. Structured Biomarker Landscape Map

Provide a structured map organized by use case first, then biomarker class.

For each biomarker entry include:

  • biomarker / signature / model name,
  • type / modality,
  • specimen / source,
  • intended use case,
  • evidence summary,
  • validation level,
  • biomarker maturity tier,
  • translation-readiness note,
  • major limitations.

D. Biomarker Maturity Layer Summary

Summarize the field using the strict maturity system from

references/biomarker-maturity-framework.md
.

At minimum, state:

  • which biomarker areas are mostly Tier 1–2,
  • which have reached Tier 3,
  • whether any area legitimately approaches Tier 4,
  • whether there is any real Tier 5 evidence,
  • where maturity is often overstated.

E. Inconsistencies, Controversies, and Failure Modes

Summarize:

  • biomarkers with conflicting reports,
  • reasons for non-reproducibility,
  • assay/platform inconsistencies,
  • endpoint-definition problems,
  • transferability concerns,
  • common overclaim patterns.

F. Validation and Translation Readiness Summary

At the field level, state:

  • which biomarker categories are mostly discovery-stage,
  • which have external validation,
  • which remain analytically or operationally weak,
  • what currently blocks translation.

G. Priority Opportunities and Under-Validated Niches

List the most important follow-up opportunities, such as:

  • biomarker classes needing external validation,
  • subtype / population gaps,
  • specimen-comparison gaps,
  • benchmark-comparison gaps,
  • assay-standardization gaps,
  • implementation-readiness gaps.

H. Primary Recommended Direction

Recommend one best next-step direction and explain:

  • why this direction is stronger than alternatives,
  • what evidence supports it,
  • what minimum next validation is required,
  • what the main failure risk is.

I. Self-Critical Risk Review

Include:

  • strongest part of the map,
  • most assumption-dependent part,
  • most likely overcalled biomarker area,
  • easiest-to-misread maturity signal,
  • likely reviewer criticism,
  • fallback interpretation if the top direction weakens under stricter validation.

J. Retrieved and Verified References

List the retrieved references used for the scan.

Reference rules:

  • do not fabricate citations, PMIDs, DOIs, trial names, or guideline status,
  • separate peer-reviewed evidence from preprints if both are used,
  • do not overstate any paper beyond what it directly supports,
  • distinguish primary studies, systematic reviews/meta-analyses, and guideline/consensus evidence whenever possible,
  • do not present unsourced field beliefs as literature-backed conclusions,
  • if evidence is thin or citation certainty is limited, say so explicitly.

Strict Biomarker Maturity Table Standard

When assigning maturity, use the following default reporting table logic.

Maturity TierWorking LabelMinimum Evidence StandardWhat It Still Cannot Claim
Tier 1Exploratory signalDiscovery-stage association only; no meaningful independent validationCannot claim robustness, reproducibility, or translational relevance
Tier 2Early validated candidateInternal validation or limited external retrospective support, but evidence remains narrowCannot claim stable generalizability or implementation readiness
Tier 3Repeatedly supported but still translationally incompleteRepeated support across independent cohorts/settings, yet key barriers remainCannot claim near-clinical readiness if assay, comparator, or operational evidence is weak
Tier 4Near-translation candidateStrong multi-cohort support plus practical assay/workflow plausibility and clearer clinical positioningCannot claim routine care adoption without prospective / implementation-grade evidence
Tier 5Clinically embedded / guideline-adjacent biomarkerFormal role in routine workflow, consensus pathway, or guideline-adjacent context clearly supportedCannot be assigned without explicit real-world clinical embedding evidence

Important rule: validation level and maturity tier are related but not identical. A biomarker may have external validation yet still remain only Tier 2 or Tier 3 if assay burden, comparator weakness, transferability, or workflow feasibility remain poor.


Formatting Expectations

  • Use a map-style output, not a long narrative review.
  • Prefer explicit labels and compact evidence statements.
  • Always distinguish use case, biomarker type, validation level, and maturity tier.
  • Do not merge diagnostic, prognostic, predictive, and monitoring claims into one row unless the evidence genuinely supports multiple roles.
  • When the field is large, group biomarkers into meaningful classes instead of generating a flat exhaustive list.
  • When evidence is uneven, show that unevenness directly instead of smoothing it into a balanced-sounding summary.

Hard Rules

  1. Never present exploratory association as biomarker maturity.
  2. Always separate diagnostic, prognostic, predictive, and monitoring claims.
  3. Always state specimen and assay context when relevant.
  4. Do not treat signatures, panels, and single markers as interchangeable.
  5. Validation level must be assigned separately from maturity tier.
  6. External validation matters more than novelty.
  7. A strong AUROC / C-index in one retrospective cohort is not biomarker maturity.
  8. When evidence conflicts, represent the conflict directly rather than averaging it away.
  9. If guideline / consensus support is absent, do not imply routine clinical adoption.
  10. If the user asks for a broad scan, prioritize structure and evidence hierarchy over completeness theater.
  11. Always include a self-critical review before the primary recommendation.
  12. Never assign Tier 4 or Tier 5 language casually; those tiers require explicit evidence beyond repeated association.
  13. Never fabricate references, PMIDs, DOIs, trial names, or validation claims.
  14. Do not present unsourced field beliefs or vague memory as literature-backed conclusions.
  15. Always distinguish exploratory reports, retrospective validation, external validation, prospective evidence, and clinical implementation evidence.
  16. Do not infer biomarker maturity from popularity, citation volume, or isolated performance metrics alone.
  17. If citation certainty is insufficient, explicitly label the point as unverified or evidence-limited instead of filling the gap.

What This Skill Should Not Do

This skill should not:

  • generate imaginary biomarker opportunities,
  • recommend patient care decisions,
  • force every biomarker into one numerical ranking,
  • confuse biological plausibility with deployable clinical value,
  • hide weak validation behind polished language,
  • pretend a sparse or contradictory field is mature.

Quality Standard

A high-quality output from this skill should read like a decision-useful biomarker evidence map.

The user should come away understanding:

  • which biomarker spaces are crowded,
  • which biomarkers are promising,
  • which are weak, inconsistent, or overclaimed,
  • what level of validation the field has actually reached,
  • what maturity tier different biomarker classes truly deserve,
  • how reliable the literature support is for the main claims,
  • and what the smartest next step would be.