Mycelium diamond-assess
Use to evaluate the current state of a diamond. Checks theory gates, confidence levels, and recommends next action.
git clone https://github.com/haabe/mycelium
T=$(mktemp -d) && git clone --depth=1 https://github.com/haabe/mycelium "$T" && mkdir -p ~/.claude/skills && cp -r "$T/.claude/skills/diamond-assess" ~/.claude/skills/haabe-mycelium-diamond-assess && rm -rf "$T"
.claude/skills/diamond-assess/SKILL.mdDiamond Assess Skill
Evaluate current diamond state and recommend next action.
Workflow
-
Identify the diamond: Which diamond (ID, scale, phase) is being assessed?
-
Gather current state:
- Current phase (Discover/Define/Develop/Deliver)
- Evidence collected so far
- Confidence score with breakdown
- Blockers or risks
-
Check theory gates for next transition:
- Reference theory-gates.md for the current transition
- Check
fromproduct_type
-- gates conditioned on product_type include:diamonds/active.yml- Security Gate: full OWASP for software/ai_tool; platform-only for content; infra-only for service
- Delivery Metrics Gate: routes to product-type-appropriate metrics canvas
- Service Quality Gate: Downe applies to consumption experience for all product types; Nielsen only for digital interfaces
- Evaluate each applicable gate: Pass / Fail / Insufficient Evidence / N/A (if gate doesn't apply to this product_type)
- Document what is missing for failed gates
-
Check confidence threshold:
- Reference confidence-thresholds.yml for the current scale
- Apply
to compute effective threshold (see confidence-thresholds.yml)project_type_adaptations - Compare current confidence to the effective threshold
- Identify what would increase confidence
-
Check for anti-patterns:
- Reference anti-patterns.md
- Flag any detected failure modes
- For L1/L2 diamonds: also check for system archetypes (Senge) — Fixes That Fail, Shifting the Burden, Limits to Growth, Eroding Goals
- At L3->L4 transitions: also run the Design Completeness Check (quality/CLAUDE.md) to verify all layers of the product design stack have evidence. Source: Mill, building on Garrett.
-
Check canvas health:
- Run the
checks inline: missing required files, stale confidence, inconsistent evidence types/canvas-health - Report any critical or warning-level findings
- This catches silent canvas degradation before it affects progression decisions
- Run the
6b. Check metric snapshot freshness (v0.14; L0/L1/L2/L5 only):
- If the current diamond scale is L0, L1, L2, or L5 AND
exists:.claude/jit-tooling/active-metrics.yml- For each
source, find the newest file instatus: active
..claude/evals/metrics/<source>/ - If the newest snapshot is >7 days old (or missing entirely), flag as a warning and recommend
./metrics-pull - If
is missing, recommendactive-metrics.yml
(softer — info-level, not a gate)./metrics-detect
- For each
- Rationale: evidence loops for Purpose/Strategy/Opportunity/Market depend on external signal freshness. A stale snapshot silently anchors confidence.
- Do NOT block progression on stale snapshots — this is a NUDGE, not a gate.
- Check corrections.md:
- Any relevant past mistakes to avoid?
7b. Check trio perspective coverage (Torres Product Trio):
- For the current diamond phase, verify all three perspectives (product/design/engineering) have been applied.
- Reference
§Trio Perspective Requirement for the per-scale coverage matrix.engine/theory-gates.md - Flag any missing perspectives as a gap: "Design perspective not yet applied at L[X]. Consider running
or/usability-check
."/service-check - If perspectives are in conflict, recommend
.engine/perspective-resolution.md
-
Coaching check (Rother's Coaching Kata): Surface these five questions in the output to prompt the human's thinking:
- What is the target condition for this diamond? (What does "done" look like?)
- What is the actual condition right now? (Summarize from steps 2-7 above)
- What obstacles are preventing progress? Which one are you addressing now?
- What is your next step? What do you expect will happen? (Force a prediction before acting)
- When can we check what we learned from that step? (Commit to a review point) The coach (human) should answer these, not the agent. The agent surfaces them. Source: Rother (Toyota Kata) — the 5 questions install scientific thinking as a daily habit.
-
Log assessment in decision-log.md (MANDATORY):
- APPEND a
entry to### Diamond Assessmentharness/decision-log.md - Include: diamond ID and scale, gates passed/failed, current confidence with rationale, evidence gaps
- This log entry is essential for auditability — every assessment should be documented
- APPEND a
-
Recommend next action:
- If all gates pass and confidence meets threshold: recommend transition to next phase
- If gates fail: recommend specific actions to address failures
- If confidence is low: recommend evidence-gathering activities
- If anti-patterns detected: recommend corrective actions
- If regression needed: recommend which phase to return to and why
-
Play devil's advocate: Before recommending progression, ask:
- What are we most likely wrong about?
- What evidence have we dismissed?
- Is there a simpler path we're overlooking?
-
Report harness thickness (informational):
- Count: total skills, active guardrails, mandatory reads, hooks, theory gates
- Current: 44 skills, 34 guardrails, 4 mandatory reads, 5 hook layers, 12 gates
- If thickness has increased since last assess, note it
- This is observability, not a gate — purely informational
- Source: Trivedy (Anatomy of an Agent Harness, LangChain blog — "scaffolding should decrease as models improve," but harnesses remain valuable as they engineer systems around model intelligence)
Output Format
ALWAYS output in plain language first, then technical details. Use
.claude/engine/status-translations.md for translations.
ALWAYS render the journey map first. Follow
.claude/engine/wayfinding.md to render the "You Are Here" map before any other output. This orients the user to where they are in the full L0→L5 progression before diving into gate details.
[Journey map from wayfinding.md — rendered first] ## Where We Are Current focus: [plain-language description from status-translations.md] [1-2 sentences of context] Confidence: [plain word] ([number], [Gilad level]) -- [why this level, what would increase it] ## Progress [N] of [M] diamonds complete: [Name]: [STATUS] -- [plain-language one-liner] [Name]: [STATUS] -- [plain-language one-liner] ## Theory Gate Check (for next transition) | Gate | Status | Suggested Skill | |------|--------|----------------| | Evidence | Pass/Fail | /user-interview or /assumption-test | | Four Risks | Pass/Fail | /assumption-test | | ... | ... | ... | ## What I'd Challenge (Devil's Advocate) - [Key assumption to question] - [Evidence gap to flag] ## Coaching Check (for the human) 1. What does "done" look like for this diamond? 2. Given what we know now, what's the biggest obstacle? 3. What's your next step -- and what do you expect will happen? 4. When should we check what we learned? ## Recommended Next Step [Plain-language recommendation with theory justification] Suggested actions: - /skill-name -- [why this is relevant now] - /skill-name -- [why this is relevant now]
Theory Citations
- Torres: Evidence-based progression
- Gilad: Confidence scoring with contextual explanation
- Cagan: Four risks assessment
- Snowden: Cynefin classification
- Shotton/Kahneman: Devil's advocate bias check
- Rother: Coaching Kata (5 questions for scientific thinking)