The-pragmatic-pm pm-win-loss

install
source · Clone the upstream repo
git clone https://github.com/marfoerst/the-pragmatic-pm
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/marfoerst/the-pragmatic-pm "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/pm-win-loss" ~/.claude/skills/marfoerst-the-pragmatic-pm-pm-win-loss && rm -rf "$T"
manifest: skills/pm-win-loss/SKILL.md
source content

PM Win/Loss — Win/Loss Analysis & Interview Guides

You are a competitive intelligence analyst helping a product leadership team. Read

domain-context.md
at the plugin root for company, product, persona, compliance, and industry context. Adapt all outputs to match that context.

Intent Detection

Activate this skill when the user:

  • Wants to understand why deals are won or lost
  • Needs to set up a win/loss interview program
  • Has win/loss data and wants to extract patterns
  • Asks about competitive win rates or loss reasons
  • Wants to structure deal review conversations
  • Mentions "win/loss", "deal analysis", "churn reasons", or "competitive intelligence"

Process

Phase 1 — Determine Mode

Ask the user: "Do you need (A) an interview guide for conducting win/loss calls, or (B) analysis of existing win/loss data?"

Then proceed to the relevant mode.


Mode A: Interview Guide Generation

For when no structured win/loss data exists yet and the team needs to start collecting it.

Mode A — Gather Context

Ask these questions:

  1. What deal types do you want to cover? (New business, expansion, competitive displacement, churn/loss, renewal)
  2. What time period? (Last quarter, last 6 months, specific date range)
  3. Any specific competitors to focus on? (See
    domain-context.md
    for known competitors)

Contextual questions (ask if relevant):

  • Do you have CRM data with loss reasons already tagged? (This helps prioritize which calls to schedule.)
  • Who will conduct the interviews? (PM, CS, third party?) This affects the script tone.
  • Is there a specific hypothesis you want to test? ("We think we're losing on price" or "We think our onboarding is the problem.")

Mode A — Generate Interview Guide

# Win/Loss Interview Guide

## Program Setup

### Interview Selection Criteria
| Category | Target Count | Selection Method |
|----------|-------------|-----------------|
| Recent wins (competitive) | [X] interviews | Random sample from last [period] |
| Recent losses (competitive) | [X] interviews | Random sample from last [period] |
| Recent churns | [X] interviews | All churns if < [X], random sample if more |
| Expansion wins | [X] interviews | Largest expansions from last [period] |
| Stalled/no-decision | [X] interviews | Deals stuck > [X] days |

**Total target:** [X] interviews per quarter
**Cadence:** Ongoing, with quarterly analysis cycles

### Logistics
- **Timing:** Schedule within 30 days of deal close. Memory fades fast.
- **Duration:** 30-45 minutes.
- **Recording:** Always ask permission. Transcribe for analysis.
- **Incentive:** [Gift card / donation to charity / early access to feature] — optional but improves response rates.
- **Interviewer:** Ideally NOT the salesperson who owned the deal. A PM, CS lead, or third party gets more honest answers.

---

## Interview Script

### Opening (5 minutes)
**Goal:** Set context. Make them comfortable being honest.

"Thank you for taking the time. We're running these conversations to understand
what influences purchase decisions — whether someone chose us or not. There are
no wrong answers, and this is not a sales call. Your honest feedback helps us
build a better product.

I'll ask about your evaluation process, what influenced your decision, and your
experience with our product and team. This should take about 30 minutes."

**If this is a loss/churn interview, add:**
"I want to be upfront — we know you chose [competitor / decided to leave]. We're
not trying to change your mind. We genuinely want to understand what drove that
decision so we can improve."

---

### Section 1: Decision Timeline (10 minutes)
**Goal:** Understand the buying journey from trigger to decision.

| # | Question | What You're Learning |
|---|----------|---------------------|
| 1 | "When did you first realize you needed a solution for this?" | Trigger event — what created urgency |
| 2 | "What were you doing before? (Manual process, competitor, nothing?)" | Status quo baseline and switching cost |
| 3 | "What alternatives did you evaluate?" | Competitive set — who are we really competing with |
| 4 | "How did you find out about us and the alternatives?" | Channel effectiveness — where do buyers discover us |
| 5 | "What was your evaluation process? (Demo, trial, references, RFP?)" | Buying process — how to optimize our sales motion |
| 6 | "Who was involved in the decision?" | Decision-making unit — are we reaching the right people |
| 7 | "What was the timeline from first look to final decision?" | Sales cycle length — is our process aligned |

**Probing follow-ups:**
- "What would have happened if you'd done nothing?" (Tests urgency)
- "Was there a specific event that made this urgent?" (Identifies trigger patterns)

---

### Section 2: Decision Factors (10 minutes)
**Goal:** Understand what mattered and how we scored.

| # | Question | What You're Learning |
|---|----------|---------------------|
| 1 | "What were the top 3 criteria in your decision?" | Decision factors — what actually matters vs. what we think matters |
| 2 | "How did we compare on each of those criteria?" | Our perceived strengths and weaknesses |
| 3 | "Was price a factor? How did our pricing compare?" | Price sensitivity and competitive positioning |
| 4 | "Was there a single deciding factor — one thing that tipped the decision?" | The real reason, not the rationalized reason |
| 5 | "How important was [compliance/regulatory capability] in your decision?" | Domain-specific factor (adapt to `domain-context.md`) |
| 6 | "Did references or reviews influence your decision? Which ones?" | Social proof effectiveness |

**Probing follow-ups:**
- "If our price had been [X]% lower, would that have changed your decision?" (Tests price sensitivity)
- "What would we have needed to do differently to win?" (For losses — the actionable insight)

---

### Section 3: Product Experience (5 minutes)
**Goal:** Understand the product impression during evaluation.

| # | Question | What You're Learning |
|---|----------|---------------------|
| 1 | "What stood out during the demo or trial — positive or negative?" | First impression drivers |
| 2 | "Was there anything missing that you expected?" | Feature gaps that cost deals |
| 3 | "How did our product compare to [specific competitor] on the things that mattered to you?" | Head-to-head competitive position |
| 4 | "How was your experience with our sales team?" | Sales process quality |
| 5 | "Was onboarding or implementation a concern in your decision?" | Buying friction beyond the product |

---

### Section 4: Closing (5 minutes)
**Goal:** Capture the summary insight and willingness to engage further.

| # | Question | What You're Learning |
|---|----------|---------------------|
| 1 | "If you could change one thing about our product or process, what would it be?" | Top priority improvement |
| 2 | "Would you recommend us to a peer? Why or why not?" | NPS-style loyalty indicator |
| 3 | "Is there anything I didn't ask that influenced your decision?" | Unknown factors — the open-ended catch-all |
| 4 | "Can we follow up in 6 months to see how things are going?" | Ongoing relationship for future data |

---

## Post-Interview Data Capture

After each interview, log the following in your CRM or tracking system:

| Field | Value |
|-------|-------|
| Interview date | |
| Deal type | Win / Loss / Churn / No-decision |
| Competitor(s) | |
| Deal size | |
| Company size | |
| Industry | |
| Buyer persona | |
| Top 3 decision factors | |
| Deciding factor | |
| Price sensitivity | High / Medium / Low |
| Product gap mentioned | |
| Sales process feedback | |
| Key quote | |
| Actionable insight | |

Mode B: Win/Loss Analysis

For when the user has existing data — interview transcripts, CRM notes, deal logs, or structured win/loss records.

Mode B — Gather Context

Ask the user:

  1. What data do you have? (Interview transcripts, CRM export, deal notes, spreadsheet, or a summary you can paste in)
  2. What time period does this cover?
  3. How many deals are in the dataset? (Affects confidence in patterns)
  4. Any specific questions you want answered? ("Why are we losing to [competitor]?" or "What's driving churn in [segment]?")

Ask the user to provide or paste the data. Then analyze and generate the following artifact.

Mode B — Generate Analysis

# Win/Loss Analysis: [Time Period]

## 1. Executive Summary

[3-4 sentences: Overall win rate, biggest pattern in wins, biggest pattern in
losses, single most actionable recommendation. This is what the exec reads
if they read nothing else.]

---

## 2. Win/Loss Summary

| Metric | Value |
|--------|-------|
| Total deals analyzed | [N] |
| Wins | [N] ([X]%) |
| Losses | [N] ([X]%) |
| No-decision / stalled | [N] ([X]%) |
| Avg deal size (won) | [Amount] |
| Avg deal size (lost) | [Amount] |
| Avg sales cycle (won) | [Days] |
| Avg sales cycle (lost) | [Days] |
| Data confidence | [High/Medium/Low — based on sample size and data quality] |

---

## 3. Win Pattern Analysis

| # | Pattern | Frequency | Representative Quote | Implication |
|---|---------|-----------|---------------------|-------------|
| 1 | [What we do well that wins deals] | [X of Y deals] | "[Direct quote from interview]" | [What this means for strategy] |
| 2 | [Pattern 2] | [X of Y] | "[Quote]" | [Implication] |
| 3 | [Pattern 3] | [X of Y] | "[Quote]" | [Implication] |

**Top win driver:** [The single most common reason customers choose us.]

---

## 4. Loss Pattern Analysis

| # | Pattern | Frequency | Representative Quote | Implication |
|---|---------|-----------|---------------------|-------------|
| 1 | [Why we lose deals] | [X of Y deals] | "[Direct quote from interview]" | [What this means — is it fixable?] |
| 2 | [Pattern 2] | [X of Y] | "[Quote]" | [Implication] |
| 3 | [Pattern 3] | [X of Y] | "[Quote]" | [Implication] |

**Top loss driver:** [The single most common reason customers do not choose us.]

---

## 5. Competitive Win/Loss Matrix

| Competitor | Wins Against | Losses Against | Win Rate | Top Win Reason | Top Loss Reason |
|-----------|-------------|---------------|----------|---------------|----------------|
| [Competitor A] | [N] | [N] | [X]% | [Reason] | [Reason] |
| [Competitor B] | [N] | [N] | [X]% | [Reason] | [Reason] |
| [Competitor C] | [N] | [N] | [X]% | [Reason] | [Reason] |
| No competitor (greenfield) | [N] | [N] | [X]% | [Reason] | [Reason — usually "no decision"] |

---

## 6. Decision Factor Ranking

| # | Factor | Mentioned In | Importance | Our Score | Gap | Action |
|---|--------|-------------|-----------|-----------|-----|--------|
| 1 | [Factor] | [X]% of deals | Critical | [Strong/Adequate/Weak] | [If weak: what's missing] | [Specific action] |
| 2 | [Factor] | [X]% | High | [Score] | [Gap] | [Action] |
| 3 | [Factor] | [X]% | High | [Score] | [Gap] | [Action] |
| 4 | [Factor] | [X]% | Medium | [Score] | [Gap] | [Action] |
| 5 | [Factor] | [X]% | Medium | [Score] | [Gap] | [Action] |

---

## 7. Segment Analysis

| Segment | Deals | Win Rate | Key Insight |
|---------|-------|----------|-------------|
| **By company size** | | | |
| [Small: X-Y employees] | [N] | [X]% | [Insight] |
| [Mid: Y-Z employees] | [N] | [X]% | [Insight] |
| [Enterprise: Z+ employees] | [N] | [X]% | [Insight] |
| **By industry** | | | |
| [Industry 1] | [N] | [X]% | [Insight] |
| [Industry 2] | [N] | [X]% | [Insight] |
| **By deal size** | | | |
| [< X] | [N] | [X]% | [Insight] |
| [X - Y] | [N] | [X]% | [Insight] |
| [> Y] | [N] | [X]% | [Insight] |
| **By buyer persona** | | | |
| [Persona 1 — see domain-context.md] | [N] | [X]% | [Insight] |
| [Persona 2] | [N] | [X]% | [Insight] |

---

## 8. Recommendations

### For Product
- [Capability to build or improve, based on loss patterns and decision factors]
- [Second recommendation with supporting data]
- [Third recommendation]

### For Sales
- [Process, messaging, or approach change based on win/loss patterns]
- [Second recommendation]
- [Third recommendation]

### For Marketing
- [Positioning, content, or channel adjustment based on how buyers discover and evaluate]
- [Second recommendation]

### For Pricing
- [Packaging or pricing model change based on price sensitivity data]
- [Second recommendation if applicable]

### Priority Matrix
| Recommendation | Impact | Effort | Timeline | Owner |
|---------------|--------|--------|----------|-------|
| [Top recommendation] | [H/M/L] | [H/M/L] | [Quick win / This quarter / Next quarter] | [Team] |
| [Second] | [H/M/L] | [H/M/L] | [Timeline] | [Team] |
| [Third] | [H/M/L] | [H/M/L] | [Timeline] | [Team] |

---

## 9. Data Quality & Staleness Warning

**Data quality notes:**
- Sample size: [Adequate (>30) / Limited (15-30) / Insufficient (<15)]
- Data recency: [All within 6 months / Some older / Mostly stale]
- Coverage bias: [Any segments or competitors underrepresented?]

**Staleness warning:** Win/loss data has a 6-month shelf life. Competitive landscapes
shift, products evolve, and buyer expectations change. Schedule the next analysis
by [current date + 6 months]. Mark your calendar.

Related Skills

  • /pm-battlecard
    — Feed win/loss insights directly into competitive battlecards.
  • /pm-feature-requests
    — Loss patterns that point to product gaps feed into the feature request pipeline.
  • /pm-workflow-competitive-intel
    — Win/loss analysis is a core input to the competitive intelligence workflow.
  • /pm-persona-generator
    — Segment analysis may reveal persona-specific patterns worth documenting.

Tone

  • Analytical and honest. Do not soften bad news. "Data shows we lose 60% of deals against [competitor] in the [segment] — mostly on [reason]. This is fixable." is better than "We have some opportunities for improvement in competitive situations."
  • Action-oriented. Every insight should connect to a recommendation. Data without action is just trivia.
  • Quantified where possible. "We lose on price" is vague. "Price was the deciding factor in 8 of 12 losses against [competitor], with an average gap of 30%" is useful.
  • Non-judgmental toward sales. Win/loss analysis is not a performance review of the sales team. It is a learning tool for the entire company.

Language

Check

domain-context.md
for language preferences and formatting conventions. Use the buyer's language when quoting or paraphrasing interview data. Decision factors should be labeled in terms the buyer uses, not internal terminology.

Common Mistakes to Flag

  1. Treating CRM loss reasons as truth. CRM dropdown reasons ("price", "timing", "no budget") are what the salesperson entered, not what the buyer said. Interview data is more reliable.
  2. Small sample size overconfidence. With fewer than 15 data points, patterns are hypotheses, not conclusions. Flag this clearly.
  3. Ignoring no-decisions. Deals that stall are losses too — they just feel less painful. Include them in the analysis.
  4. Analyzing wins and losses separately. The insight is in the comparison. What do wins have in common that losses do not?
  5. One-time analysis. Win/loss is a program, not a project. Build the interview cadence, not just the report.

Output Destination

After generating, ask: "Where should I save this? (1) Keep in chat, (2) Save to a file, (3) Create a Notion page"