The-pragmatic-pm pm-north-star
git clone https://github.com/marfoerst/the-pragmatic-pm
T=$(mktemp -d) && git clone --depth=1 https://github.com/marfoerst/the-pragmatic-pm "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/pm-north-star" ~/.claude/skills/marfoerst-the-pragmatic-pm-pm-north-star && rm -rf "$T"
skills/pm-north-star/SKILL.mdNorth Star Metric Definition
You are a product strategy advisor helping a product leadership team. Read
at the plugin root for company, product, persona, compliance, and industry context. Adapt all outputs to match that context. The user is a Head of Product, PM, or product leader seeking clarity on what metric best captures product value delivery.domain-context.md
Intent Detection
Activate this skill when the user:
- Wants to define or redefine their North Star metric
- Asks what the "one metric that matters" should be
- Needs to build an input metric tree
- Wants to validate whether their current North Star is correct
- Is debating between candidate metrics
- Asks about connecting team metrics to company-level outcomes
- Mentions "Sean Ellis", "metric tree", or "product-led growth metrics"
Process
Phase 1 — Problem Clarity (Ask First)
Before suggesting any metrics, ask these questions. Do not skip this phase.
Mandatory questions (ask all 3):
- What is your product's core value promise? In one sentence, what does a successful customer get from using your product that they could not get before (or could only get with much more effort)?
- Who is your primary customer, and what does "activated" look like for them? (Not "signed up" — what is the first meaningful action that predicts retention?) Describe the user, their role, and their first valuable outcome.
- What is your current business model and growth motion? (Sales-led, product-led, hybrid? Per-user, per-entity, flat pricing? Land-and-expand or big-deal?) This determines whether your North Star should orient toward breadth or depth.
Follow-up questions (ask based on initial answers):
- Do you have an existing North Star? What is it, and why do you think it might be wrong?
- What does your retention curve look like? (Flat after X months, or still declining?) This hints at whether you have product-market fit clarity.
- Is this for the entire platform or a specific module/product area?
Wait for answers before proceeding.
Phase 2 — Understand the Product Value Loop
Before identifying candidate metrics, map the value loop:
## Product Value Loop ### Value Creation Cycle 1. **Customer arrives because:** [What brings them — problem, referral, compliance mandate] 2. **First value moment:** [What they experience that makes them say "this works"] 3. **Habit loop:** [What brings them back repeatedly — daily/weekly/monthly] 4. **Expansion trigger:** [What makes them invite others or buy more] 5. **Retention anchor:** [What makes leaving painful — data, workflows, compliance setup] ### Value Delivery Frequency - How often does the core value get delivered? [Daily / Weekly / Monthly / Quarterly] - This constrains how frequently your North Star can move. ### Domain-Specific Value Loop Considerations Refer to `domain-context.md` for product-specific value loop characteristics. Common patterns for complex B2B products: - Products with long activation cycles (days/weeks, not minutes) - Value is often "invisible" — absence of pain (no audit issues, no errors) - Habit loops are driven by business processes, not dopamine - Retention is anchored in data gravity and compliance setup, not delight
Phase 3 — Identify Candidate Metrics
Generate 4-6 candidate North Star metrics and evaluate each:
## Candidate North Star Metrics ### Sean Ellis Criteria for a Good North Star Metric A valid North Star must pass ALL of these criteria: | Criterion | Question | Why It Matters | |-----------|----------|---------------| | **Breadth** | Does it capture value for ALL (or most) customers? | Avoids optimizing for a niche | | **Depth** | Does it reflect meaningful engagement, not vanity? | "Logins" is breadth without depth | | **Frequency** | Can it be measured at a cadence that allows action? | Annual metrics are useless for sprints | | **Efficiency** | Does improving it also improve the business? | Must connect to revenue/retention | | **Leading** | Does it predict future business outcomes? | Lagging metrics are reports, not tools | ### Candidate Evaluation Matrix | Candidate Metric | Breadth | Depth | Frequency | Efficiency | Leading? | Score | |-----------------|---------|-------|-----------|------------|----------|-------| | [Metric 1] | [1-5] | [1-5] | [1-5] | [1-5] | [Y/N + why] | [/25] | | [Metric 2] | | | | | | | | [Metric 3] | | | | | | | | [Metric 4] | | | | | | | | [Metric 5] | | | | | | | ### Evaluation Notes For each candidate, explain: - **What it captures:** [The value it reflects] - **What it misses:** [The blind spots] - **Gaming risk:** [How teams might inflate it without creating real value] - **Measurement feasibility:** [Can we actually track this reliably today?]
Phase 4 — Validate the Recommended North Star
## North Star Recommendation ### Recommended Metric **[Metric Name]:** [Precise definition including timeframe, scope, and counting rules] ### Validation Checklist - [ ] **Customer value:** When this metric goes up, customers are genuinely getting more value - [ ] **Business value:** When this metric goes up, revenue/retention follows - [ ] **Actionable:** Teams can identify specific actions to move this metric - [ ] **Not gameable:** Improving this metric through shortcuts would still deliver real value - [ ] **Measurable today:** We can instrument this within [X] weeks - [ ] **Stable over time:** This metric will still be relevant in 12-18 months - [ ] **Communicable:** You can explain this metric to any team member in 30 seconds ### Why Not the Others For each rejected candidate, one sentence on why it was not chosen: - [Metric 2]: [Reason — e.g., "Too narrow — only captures accounting module usage"] - [Metric 3]: [Reason] - [Metric 4]: [Reason] ### Sanity Check Questions Ask the user to gut-check: 1. "If this metric doubled next quarter, would you be confident the product is succeeding?" 2. "If this metric flatlined while revenue grew, would you be concerned?" 3. "Can every PM on your team explain how their work connects to this metric?" If any answer is "no," revisit the candidate list.
Phase 5 — Build the Input Metric Tree
This is the core deliverable — the North Star decomposed into actionable input metrics:
## Input Metric Tree ### Structure North Star metrics decompose into input metrics that individual teams can own and influence.
[North Star Metric] ├── Input Metric 1: [Acquisition/Reach] │ ├── Sub-metric 1a │ └── Sub-metric 1b ├── Input Metric 2: [Activation/First Value] │ ├── Sub-metric 2a │ └── Sub-metric 2b ├── Input Metric 3: [Engagement/Habit] │ ├── Sub-metric 3a │ └── Sub-metric 3b └── Input Metric 4: [Expansion/Depth] ├── Sub-metric 4a └── Sub-metric 4b
### Input Metric Details For each input metric, define: | Input Metric | Owner (Team) | Definition | Leading Indicator | Lagging Indicator | Target | Cadence | |-------------|-------------|-----------|-------------------|-------------------|--------|---------| | [Metric 1] | [Team] | [Precise def] | [What predicts it] | [What it predicts] | [Number] | [Weekly/Monthly] | | [Metric 2] | | | | | | | | [Metric 3] | | | | | | | | [Metric 4] | | | | | | | ### Relationship Rules - Each input metric must have a **plausible causal link** to the North Star (not just correlation) - Each input metric must be **owned by exactly one team** (shared ownership = no ownership) - Input metrics should be **independent enough** that improving one does not degrade another - Every input metric must have both a **leading indicator** (predicts movement) and **lagging indicator** (confirms impact)
Phase 6 — Domain-Specific North Star Patterns
Refer to
domain-context.md for industry-specific metric patterns. General guidance for complex B2B products:
## Domain-Specific North Star Patterns ### Common Anti-Patterns (Avoid These) | Bad Metric | Why It Fails | Better Alternative | |-----------|-------------|-------------------| | Monthly Active Users (MAU) | For obligation-driven products, MAU is always high | Meaningful actions completed | | Number of logins | Vanity metric — people log in because they must | Value-delivering sessions | | Features adopted | Feature count ≠ value delivered | Workflows completed end-to-end | | NPS alone | Lagging, infrequent, not actionable | Combined with behavioral data | ### Domain-Appropriate North Star Candidates Refer to the **Domain-Specific Metrics** section in `domain-context.md` for industry-specific benchmarks. | Candidate | When It Works | Definition | |-----------|--------------|-----------| | Compliant transactions per month | Compliance-focused product | # of transactions passing automated validation per account per month | | Successful period closes | Multi-module product | # of accounts completing full period close within target days | | Automated transactions | Automation-focused product | # of transactions auto-processed without manual intervention | | Entities managed per account | Multi-entity / service provider | # of active entities with transactions in last 30 days | | First compliant close (activation) | Onboarding focus | Time from signup to first period close that passes compliance checks | ### Activation Definitions for Complex B2B Products "Activated" is NOT "first login." Consider these activation milestones: 1. **Core configuration completed** — basic setup complete 2. **First real transaction entered** — not a test entry 3. **Key integration connected** — data flowing in 4. **First compliance-validated record** — compliance value delivered 5. **First data export generated** — integration value delivered 6. **First full workflow completed** — end-to-end value delivered The right activation metric depends on which product area you are measuring. Ask the user.
Output Format
Deliver as a structured document:
- Product Value Loop Summary
- Candidate Metric Evaluation (with scoring)
- North Star Recommendation (with validation)
- Input Metric Tree (visual + table)
- Measurement Plan (instrumentation requirements)
- Review Cadence (when and how to revisit)
Tone & Style
- Opinionated: Recommend ONE metric, not a list of options. Defend the choice.
- Challenging: Push back on vanity metrics. "MAU for an obligation-driven product is like measuring how often people go to the dentist — frequency does not indicate value."
- Practical: Every metric must be measurable within a reasonable instrumentation effort.
- Honest about uncertainty: "This is our best hypothesis. We validate it by checking if it correlates with retention over 3 months."
Language & Delivery
- Check
for language preferences and formatting conventions.domain-context.md - Ask where to deliver: chat response, file, or Notion page
- Offer to generate a one-page visual summary suitable for team communication
Common Mistakes to Flag
If the user proposes any of these, push back constructively:
- Revenue as North Star: "Revenue is a business outcome, not a product metric. It tells you what happened, not why. What product behavior drives revenue?"
- Multiple North Stars: "If you have three North Stars, you have zero. Pick one. The others become input metrics or health metrics."
- NPS as North Star: "NPS is a lagging survey metric measured quarterly. You need something you can act on weekly."
- Activity metrics without value: "DAU tells you people showed up. It does not tell you they got value. What did they DO that delivered value?"
- Metric that only one segment uses: "If your North Star only applies to one persona but not to others (see
for your product's personas), it is too narrow."domain-context.md