Claude-skill-registry goal-setting
Use when establishing targets for defined metrics - validates metric is movable by team, assesses movement needed for company goals, evaluates aggressive vs conservative targets with trade-off analysis
git clone https://github.com/majiayu000/claude-skill-registry
T=$(mktemp -d) && git clone --depth=1 https://github.com/majiayu000/claude-skill-registry "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/data/goal-setting" ~/.claude/skills/majiayu000-claude-skill-registry-goal-setting && rm -rf "$T"
skills/data/goal-setting/SKILL.mdGoal Setting Workflow
Purpose
Establish ambitious but realistic targets for defined metrics by understanding company-level impact needed, validating team's ability to move the metric, and evaluating trade-offs between aggressive and conservative goals.
When to Use This Workflow
Use this workflow when:
- After defining success metrics, need to set OKRs or targets
- Planning roadmap and need to estimate impact of initiatives
- Leadership asks "what's the goal for this metric this quarter?"
- Evaluating whether proposed project is worth the investment
- Setting team or product area goals
- Translating company-level goals to product-level targets
Skills Sequence
This workflow orchestrates 3 core skills:
1. North Star Alignment ↓ (How much movement matters to company goals?) 2. Proxy Metric Selection ↓ (Validate metric is movable by this team) 3. Trade-off Evaluation ↓ (Cost of aggressive vs conservative targets) OUTPUT: Target value or % improvement, timeframe, confidence level, dependencies and assumptions
Required Inputs
Gather this information before starting:
Metric Information
-
The metric and its current baseline
- Metric name and formula
- Current value
- Example: "Weekly Active Users = 100,000"
-
Historical trends
- How has metric moved historically?
- Seasonality patterns?
- Growth rate trajectory?
- Example: "Growing 3% month-over-month for past year"
Context
-
Planned initiatives
- What are you building to move this metric?
- Estimated impact of each initiative
- Timeline for delivery
-
Comparison benchmarks (if available)
- Competitor metrics
- Industry standards
- Similar products' performance
Company Goals
-
Company-level targets
- What are company OKRs?
- How does this metric contribute?
- What movement is needed for company goals?
-
Resource constraints
- Team size and capacity
- Budget limitations
- Technical constraints
Workflow Steps
Step 1: Assess Company-Level Impact Needed (15 minutes)
Use the
skillnorth-star-alignment
Understand how much metric movement matters to company goals:
Activities:
-
Identify North Star connection
- How does your metric connect to company North Star?
- What's the multiplier or contribution factor?
- Example: "10% increase in WAU → 2% increase in company MAU"
-
Calculate company goal contribution
- If company wants X% growth in North Star
- How much must your metric grow?
- Example: "Company wants 10% MAU growth; our product = 20% of MAU; need 10% WAU growth minimum"
-
Evaluate strategic importance
- Is this metric critical path to company goals?
- Or nice-to-have but not essential?
- Priority: Critical / Important / Opportunistic
Output:
## Company-Level Impact Assessment **North Star Connection:** - Company North Star: [Metric name] - Your metric's contribution: [How it ladders up] - Multiplier/factor: [Quantify relationship] **Company Goal Translation:** - Company target: [X% growth in North Star] - Required from your product: [Y% growth in your metric] - Rationale: [Calculation or logic] **Strategic Priority:** - [Critical / Important / Opportunistic] - Why: [Explanation] **Minimum viable target:** - To meet company needs: [Minimum % or absolute value]
Step 2: Validate Metric Movability (20 minutes)
Use the
skillproxy-metric-selection
Confirm the team can actually influence this metric:
Activities:
-
Identify levers team controls
- What features/changes can team implement?
- What's outside team's control?
- What dependencies exist?
-
Estimate initiative impacts
- For each planned initiative, estimate impact
- Use historical data, A/B tests, competitor analysis
- Range estimates (low/medium/high confidence)
-
Sum potential impact
- Add up all initiatives
- Account for overlap/cannibalization
- Reality-check against historical growth rates
-
Assess confidence
- High confidence: Proven tactics, similar to past successes
- Medium confidence: Reasonable hypotheses, some validation
- Low confidence: Exploratory, unproven approaches
Output:
## Movability Assessment **Team Control:** - Direct levers: [What team can change] - Indirect influence: [What team can affect partially] - Outside control: [What team cannot affect] **Planned Initiatives:** | Initiative | Est. Impact (Low) | Est. Impact (High) | Confidence | Timeline | |------------|-------------------|---------------------|------------|----------| | [Feature A] | +2% | +5% | Medium | Q1 | | [Feature B] | +1% | +3% | High | Q1 | | [Feature C] | +3% | +8% | Low | Q2 | **Total Potential Impact:** - Conservative (low estimates): [Sum] - Aggressive (high estimates): [Sum] - Reality-checked: [Adjusted for overlap, feasibility] **Confidence Level:** - [High / Medium / Low] - Based on: [Historical performance, validation, team track record] **Dependencies:** - [Other teams, external factors, resources needed]
Step 3: Evaluate Target Trade-offs (20 minutes)
Use the
skilltradeoff-evaluation
Assess costs and benefits of aggressive vs. conservative targets:
Framework: Three target levels
Level 1: Conservative Target
- Definition: Achievable with high confidence
- Characteristics:
- Based on proven tactics
- Requires minimal risk-taking
- Accounts for setbacks
- Pro: Likely to hit target, builds credibility
- Con: May leave growth on table, less inspiring
Level 2: Ambitious Target
- Definition: Stretch goal requiring execution excellence
- Characteristics:
- Requires most initiatives to succeed
- Some unknowns but reasonable
- Pushes team but achievable
- Pro: Motivating, meaningful if hit
- Con: 50-70% confidence, some risk
Level 3: Moonshot Target
- Definition: Requires everything to go right + surprises
- Characteristics:
- All initiatives succeed maximally
- Plus unexpected wins
- Low probability but transformative
- Pro: Inspires big thinking
- Con: High failure risk, may demotivate if unrealistic
Evaluate each level:
Questions:
- Resource cost: What resources required for each level?
- Opportunity cost: What are we NOT doing to hit this?
- Risk: What breaks if we miss?
- Motivation: Does target inspire or overwhelm team?
- Stakeholder expectation: What's leadership expecting?
Output:
## Target Trade-off Analysis **Conservative Target:** [Value or % improvement] - Confidence: 90%+ - Based on: [Proven tactics, historical trends] - Pros: [List] - Cons: [List] - Resource needs: [Description] **Ambitious Target:** [Value or % improvement] - Confidence: 60-70% - Based on: [All planned initiatives + reasonable execution] - Pros: [List] - Cons: [List] - Resource needs: [Description] - Risk if missed: [Impact] **Moonshot Target:** [Value or % improvement] - Confidence: 20-30% - Based on: [Everything + unexpected wins] - Pros: [List] - Cons: [List] - Resource needs: [Description] - Risk if missed: [Impact] **Recommendation:** [Which level to commit to] - Rationale: [Why this level makes sense] - Publicly commit to: [May be different from internal stretch] - Internal stretch: [What team pushes for privately]
Step 4: Set Final Goal (10 minutes)
Synthesize analyses into clear goal statement:
Components of good goal:
- Metric: Precise definition
- Target: Specific value or % improvement
- Timeframe: Clear deadline
- Confidence: Commitment vs. aspirational
- Dependencies: What must be true
- Milestones: Intermediate checkpoints
Goal template:
## Final Goal Statement **Primary Goal:** Increase [Metric name] from [Current baseline] to [Target value] by [Date] **This represents:** [X% improvement] **Commitment level:** [Committed / Aspirational / Stretch] **Confidence:** [%] **Why this target:** 1. [Company contribution reason] 2. [Team capability reason] 3. [Strategic timing reason] **Key Initiatives:** 1. [Initiative name]: [Expected contribution] 2. [Initiative name]: [Expected contribution] 3. [Initiative name]: [Expected contribution] **Dependencies:** - [External dependency 1] - [External dependency 2] - [Resource dependency] **Milestones:** - Month 1: [Intermediate target] - Month 2: [Intermediate target] - Month 3: [Intermediate target] **Success criteria:** - Hit target: [What constitutes success] - Partial success: [What's acceptable] - Failure: [What triggers re-evaluation] **Review cadence:** - [Weekly / Bi-weekly / Monthly] check-ins - [What triggers goal revision]
Goal-Setting Frameworks
Framework 1: OKR Structure
Objective: Qualitative, inspiring Key Results: Quantitative, measurable
Example:
- Objective: Make our product indispensable for power users
- Key Result 1: Increase WAU from 100K to 120K (20% growth)
- Key Result 2: Increase daily actives from 40K to 50K (25% growth)
- Key Result 3: Improve week-2 retention from 60% to 70%
Scoring: 0-1 scale
- 0.7-1.0 = Success (ambitious targets)
- 0.4-0.6 = Partial progress
- <0.4 = Missed
Framework 2: Committed vs. Aspirational
Committed goals:
- 90%+ confidence
- Resources committed
- What you'll be evaluated on
- Example: "Grow WAU 10%"
Aspirational goals:
- 50-60% confidence
- Stretch targets
- Not penalized for missing
- Example: "Grow WAU 25%"
Framework 3: Input vs. Output Goals
Output goals (results):
- "Increase WAU 20%"
- What you want to achieve
- Ties to company goals
Input goals (activities):
- "Launch 3 major features"
- What you'll do
- Ties to team capacity
Best practice: Set both
- Primary: Output goal
- Supporting: Input goals that should deliver output
Framework 4: Leading Indicator Targets
Why: Leading indicators move before lagging outcomes
Example:
- Lagging: Retention rate (know months later)
- Leading: Activation rate (know immediately)
Set targets for:
- Leading indicators (drive focus)
- Lagging indicators (confirm success)
Common Pitfalls
| Pitfall | How to Avoid |
|---|---|
| Sandbagging (too easy) | Push to ambitious level, not conservative |
| Moonshot without plan | Ensure initiatives exist to support target |
| No company alignment | Start with North Star connection |
| Ignoring historical trends | Reality-check against past performance |
| No intermediate milestones | Break into monthly/quarterly checkpoints |
| Not considering seasonality | Adjust for known seasonal effects |
| Resource mismatch | Validate team capacity to execute |
Success Criteria
Goal setting succeeds when:
- Target value/% clearly stated
- Timeframe specific
- Company-level impact articulated
- Team's ability to move metric validated
- Initiatives mapped to target with estimated impacts
- Trade-offs between aggressive/conservative evaluated
- Confidence level honest
- Dependencies and assumptions documented
- Intermediate milestones defined
- Review cadence established
- Stakeholders aligned on goal
Real-World Example: Uber Driver Quality Program
Context
- Metric: Hours driven in 4.8+ rating bucket
- Current: 4.8M hours / 60% of total
- Timeline: Next quarter (Q1)
Step 1: Company Impact (15 min)
Company Goal: Grow Monthly Active Drivers 15% in 2025 Our Contribution: Driver quality → retention → supply growth Math: - If quality drivers stay 20% longer - And recruit 10% more drivers via referrals (quality attracts quality) - Contributes ~5 percentage points to 15% MAD goal Strategic Priority: Critical (supply is bottleneck for growth) Minimum viable: 65% of hours in 4.8+ bucket (5 percentage point improvement)
Step 2: Movability Assessment (20 min)
Team Control: - Direct: Quality dashboard, driver tips, rating transparency - Indirect: Driver training, incentives (ops team) - Outside: Rider behavior, market conditions Planned Initiatives: | Initiative | Low Impact | High Impact | Confidence | Timeline | |------------|------------|-------------|------------|----------| | Enhanced dashboard | +1% | +2% | High | Month 1 | | Quality tips program | +2% | +4% | Medium | Month 1-2 | | Referral program | +1% | +3% | Medium | Month 2-3 | | Driver coaching | +1% | +2% | High | Ongoing | Total Potential: - Conservative: +5 percentage points (65% total) - Aggressive: +11 percentage points (71% total) - Reality-checked: +7 percentage points (67% total) (Accounting for overlap, execution risk) Confidence: Medium-High (70%) - Based on: Similar programs in other markets worked - Risk: Driver adoption lower than expected Dependencies: - Ops team for coaching rollout - Data team for dashboard enhancements
Step 3: Trade-off Analysis (20 min)
Conservative Target: 65% (5 points) - Confidence: 90% - Based on: Enhanced dashboard + quality tips (proven tactics) - Pro: Highly achievable, builds credibility - Con: Doesn't capitalize on full potential, uninspiring - Resource: 2 engineers, 1 designer, 1 PM Ambitious Target: 67% (7 points) - Confidence: 70% - Based on: All initiatives, reasonable execution - Pro: Meaningful improvement, motivating team - Con: Requires referral program success (moderate risk) - Resource: +1 growth engineer for referrals Moonshot Target: 70% (10 points) - Confidence: 30% - Based on: Everything working + viral referral growth - Pro: Transformative if achieved - Con: Very low probability, may demotivate if unrealistic - Resource: Full team focus, opportunity cost high Recommendation: Ambitious (67%) - Company needs meaningful contribution (5% minimum) - Team has clear plan for 7% (all initiatives) - 70% confidence acceptable for OKR - Publicly commit to 67%, internal stretch to 69%
Step 4: Final Goal (10 min)
PRIMARY GOAL: Increase hours driven in 4.8+ rating bucket from 60% to 67% by end of Q1 2025 This represents: 7 percentage point improvement Commitment level: Aspirational OKR (targeting 0.7-1.0 score) Confidence: 70% Why this target: 1. Company contribution: Supports 15% MAD growth goal (critical path) 2. Team capability: Have clear initiatives totaling 7-11% potential 3. Strategic timing: Quality program momentum from Q4, capitalize now Key Initiatives: 1. Enhanced dashboard (Month 1): +1-2 points 2. Quality tips program (Month 1-2): +2-4 points 3. Referral program (Month 2-3): +1-3 points 4. Driver coaching (Ongoing): +1-2 points Dependencies: - Ops team partnership for coaching - Data team dashboard platform (committed) - Referral incentive budget approved Milestones: - Month 1: 62% (dashboard launch, tips program start) - Month 2: 64% (tips program maturing, referrals launching) - Month 3: 67% (all initiatives running, compounding effects) Success criteria: - Hit 67%: Full success - 64-66%: Partial success (0.5-0.7 OKR score) - <64%: Miss (requires post-mortem) Review cadence: - Weekly: Quality distribution trends - Bi-weekly: Initiative progress review - Month-end: Milestone assessment - Trigger for revision: <1 point improvement per month
Time to complete: 65 minutes
Goal Calibration Examples
Example 1: Too Conservative
Problem:
- Current: 100K WAU, growing 3% monthly
- Target: 103K in 3 months (3% growth)
- Historical trend: Would hit this doing nothing
Fix:
- Ambitious target: 115K (15% growth = 5% monthly)
- Requires initiatives beyond business-as-usual
- Stretches team but achievable
Example 2: Too Aggressive
Problem:
- Current: 100K WAU, growing 3% monthly
- Target: 200K in 3 months (100% growth)
- No clear path to doubling in 3 months
Fix:
- Reality-check: What's maximum with all initiatives?
- Maybe 125K (25% growth) if everything works
- Set 120K committed, 130K stretch
Example 3: Well-Calibrated
Problem:
- Current: 100K WAU, growing 3% monthly
- Business-as-usual trajectory: 109K in 3 months
- Target: 120K (10% growth above baseline)
Good because:
- Requires planned initiatives to succeed
- Achievable with good execution
- Meaningful above baseline
- Team bought in
Related Skills
This workflow orchestrates these skills:
- north-star-alignment (Step 1)
- proxy-metric-selection (Step 2)
- tradeoff-evaluation (Step 3)
Related Workflows
- metrics-definition: Defines metrics before setting goals
- dashboard-design: Uses goals to set alert thresholds
Time Estimate
Total: 65-75 minutes
- Step 1 (Company impact): 15 min
- Step 2 (Movability): 20 min
- Step 3 (Trade-offs): 20 min
- Step 4 (Final goal): 10 min