Claude-skill-registry-data measure-experiment-results

<!-- PM-Skills | https://github.com/product-on-purpose/pm-skills | Apache 2.0 -->

install
source · Clone the upstream repo
git clone https://github.com/majiayu000/claude-skill-registry-data
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/majiayu000/claude-skill-registry-data "$T" && mkdir -p ~/.claude/skills && cp -r "$T/data/measure-experiment-results" ~/.claude/skills/majiayu000-claude-skill-registry-data-measure-experiment-results && rm -rf "$T"
manifest: data/measure-experiment-results/SKILL.md
source content
<!-- PM-Skills | https://github.com/product-on-purpose/pm-skills | Apache 2.0 -->

name: measure-experiment-results description: Documents the results of a completed experiment or A/B test with statistical analysis, learnings, and recommendations. Use after experiments conclude to communicate findings, inform decisions, and build organizational knowledge. phase: measure version: "2.0.0" updated: 2026-01-26 license: Apache-2.0 metadata: category: reflection frameworks: [triple-diamond, lean-startup, design-thinking] author: product-on-purpose

Experiment Results

An experiment results document captures what happened when you tested a hypothesis, including statistical outcomes, segment analysis, learnings, and clear recommendations. Good results documentation turns individual experiments into organizational knowledge that improves future decision-making.

When to Use

  • After an A/B test or experiment reaches statistical significance
  • When an experiment is ended early (for any reason)
  • To communicate findings to stakeholders who weren't involved
  • During decision-making about whether to ship, iterate, or kill a feature
  • To build a repository of learnings that inform future experiments

Instructions

When asked to document experiment results, follow these steps:

  1. Summarize the Experiment Provide context: what was tested, when it ran, how much traffic it received. Link to the original experiment design document if one exists.

  2. Restate the Hypothesis Remind readers what you believed would happen and why. This frames the results interpretation.

  3. Present Primary Results Show the primary metric outcome clearly: what were the values for control and treatment? Include statistical significance (p-value), confidence intervals, and sample sizes. Be honest about whether results are conclusive.

  4. Analyze Secondary Metrics Present guardrail metrics that ensure you didn't cause unintended harm. Note any secondary metrics that moved unexpectedly—both positive and negative.

  5. Segment the Data Look for differential effects across user segments (platform, tenure, plan type, etc.). Sometimes overall results mask important segment-level insights.

  6. Extract Learnings What did you learn beyond the numbers? Include surprising findings, questions raised, and implications for the product hypothesis. Negative results are valuable learnings.

  7. Make a Recommendation Be clear: should we ship, iterate, or kill? Support the recommendation with the evidence. If the decision is nuanced, explain the trade-offs.

  8. Define Next Steps Specify what happens now—engineering work to ship, follow-up experiments, metrics to continue monitoring, or documentation to update.

Output Format

Use the template in

references/TEMPLATE.md
to structure the output.

Quality Checklist

Before finalizing, verify:

  • Statistical methods and significance are clearly stated
  • Confidence intervals are included (not just p-values)
  • Segment analysis checked for differential effects
  • Secondary/guardrail metrics are reported
  • Learnings go beyond just the numbers
  • Recommendation is clear and actionable
  • Negative or inconclusive results are reported honestly

Examples

See

references/EXAMPLE.md
for a completed example.