Pm-claude-skills rice-prioritisation
Score and rank product initiatives using the RICE framework. Use when asked to prioritise features, rank a backlog using RICE, score initiatives for quarterly planning, or apply an objective framework to a list of competing ideas. Produces a ranked RICE table with scores, quick wins and moonshot flags, dependency notes, and a recommended sequencing order.
install
source · Clone the upstream repo
git clone https://github.com/mohitagw15856/pm-claude-skills
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/mohitagw15856/pm-claude-skills "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/rice-prioritisation" ~/.claude/skills/mohitagw15856-pm-claude-skills-rice-prioritisation-9fdbb0 && rm -rf "$T"
manifest:
skills/rice-prioritisation/SKILL.mdsource content
RICE Prioritisation Skill
Apply consistent, criteria-based RICE scoring to a list of features or initiatives to produce an objective prioritisation ranking.
Required Inputs
Ask the user for these if not provided:
- List of initiatives or features to score (names and brief descriptions)
- Reach estimates (users affected per quarter — from analytics if available)
- Impact estimates (use the standard scale below)
- Effort estimates (person-months — from engineering if available)
- Quarter or planning period
RICE Definitions (adapt to your context)
- Reach: Number of users affected per quarter (use actual DAU/MAU data where available)
- Impact: Effect on your primary metric — use scale: 3=massive, 2=high, 1=medium, 0.5=low, 0.25=minimal
- Confidence: How certain are we about R and I estimates? 100%=high, 80%=medium, 50%=low
- Effort: Person-months required across all functions
RICE Formula
RICE Score = (Reach × Impact × Confidence) / Effort
Process
- For each initiative provided, gather or estimate R, I, C, E values
- Flag where estimates are weak and note what data would improve them
- Calculate RICE score for each
- Rank highest to lowest
- Flag any "quick wins" (high RICE score, low effort) and "moonshots" (high impact, high effort)
- Note dependencies between items that affect sequencing
- Validate — Cross-check: if the top-ranked item surprises the team, investigate whether an estimate is inflated. RICE is a tool, not a verdict.
Output Structure
RICE Prioritisation: [Backlog/Quarter]
| Initiative | Reach | Impact | Confidence | Effort | RICE Score | Notes |
|---|---|---|---|---|---|---|
| [name] | [n] | [score] | [%] | [months] | [score] | [flags] |
Recommended Sequence
[Top 5 initiatives with rationale]
Quick Wins (high score, low effort)
[Items to pick up alongside bigger bets]
Data Gaps to Address
[What information would most improve scoring accuracy]
Quality Checks
- Every initiative has all four RICE components estimated (even roughly)
- Confidence is 50% for anything without data backing (not 100% as a default)
- Quick wins and moonshots are explicitly called out
- Dependencies that affect sequencing are noted
- Any surprising ranking is investigated before accepting it