install
source · Clone the upstream repo
git clone https://github.com/Runchuan-BU/BioClaw
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/Runchuan-BU/BioClaw "$T" && mkdir -p ~/.claude/skills && cp -r "$T/container/skills/bio-metric-system" ~/.claude/skills/runchuan-bu-bioclaw-bio-metric-system && rm -rf "$T"
manifest:
container/skills/bio-metric-system/SKILL.mdsource content
bio-metric-system
Step 4: Metric system design (评价指标体系构建)
Build a defensible set of quantitative and qualitative metrics by extracting them from related work or adapting them from adjacent fields.
Purpose
- Extract evaluation metrics from existing literature
- Borrow metrics from adjacent domains when needed
- Organize metrics into quantitative and qualitative groups
- Explain what each metric measures and how it should be computed
Input Format
topic: [research topic] paper_count: [number of related papers] task_system: [task system from Step 2]
Workflow
Step 4.1: Extract metrics from existing work
If
paper_count >= 5, review the Results / Benchmark sections of the strongest related papers and extract:
- metric name
- what it evaluates
- formula or computation method
- expected range
- how often it appears in the field
Step 4.2: Borrow metrics from adjacent domains
If the literature is still thin, adapt metrics from a nearby field.
Examples:
- clustering agreement -> ARI / NMI
- modality agreement -> Pearson / cosine similarity
- reconstruction / registration -> MSE / MAE
- biological relevance -> marker recovery / enrichment scores
Step 4.3: Organize the metric system
Split metrics into:
- Quantitative metrics
- integration quality
- modality consistency
- registration / alignment quality
- biological agreement
- Qualitative metrics
- spatial plots
- feature plots
- latent visualizations
- heatmaps
- pathway / enrichment figures
Step 4.4: Standardize each metric
For each metric, define:
- English name
- optional Chinese reference in parentheses
- category
- what it measures
- formula (if needed)
- range / interpretation
- software implementation
- task relevance
- mapped figure / panel
Output Format
# Metric System ## Metric Sources - Extracted from related papers: - Borrowed from adjacent domains: ## Quantitative Metrics ### ARI (Adjusted Rand Index) - Category: - What it measures: - Formula: - Range: - Interpretation: - Implementation: - Relevant tasks: - Figure mapping: ### NMI (Normalized Mutual Information) - Category: - What it measures: - Formula: - Range: - Interpretation: - Implementation: - Relevant tasks: - Figure mapping: ### Pearson correlation - Category: - What it measures: - Formula: - Range: - Interpretation: - Implementation: - Relevant tasks: - Figure mapping: ## Qualitative Metrics / Visual Readouts - spatial domain map - feature plot - violin plot - UMAP / latent visualization - heatmap - pathway enrichment figure ## Next Step - Use the metric system to build the analysis system in Step 5
Recommended Core Metrics
For most manuscript-planning runs, include at least:
- ARI
- NMI
- Macro-F1 or annotation accuracy
- Pearson / cosine similarity when cross-modal agreement matters
- MSE / MAE when reconstruction or alignment quality matters
- at least one biological validation readout
Usage
/bio-metric-system "spatial multi-omics integration | paper_count: 5 | task_system: [task system from Step 2]"
Notes
- Do not overload the paper with too many metrics; prefer a compact but defendable set.
- Match each metric to a specific task claim.
- Include at least one metric that reflects biological value, not just technical fit.