Claude-skill-registry claimify
Extract and structure claims from discourse into analyzable argument maps with logical relationships and assumptions. Use when analyzing arguments, red-teaming reasoning, synthesizing debates, or transforming conversations into structured claim networks. Triggers include "what are the claims," "analyze this argument," "map the logic," or "find contradictions."
git clone https://github.com/majiayu000/claude-skill-registry
T=$(mktemp -d) && git clone --depth=1 https://github.com/majiayu000/claude-skill-registry "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/data/claimify" ~/.claude/skills/majiayu000-claude-skill-registry-claimify && rm -rf "$T"
skills/data/claimify/SKILL.mdClaimify
Extract claims from text and map their logical relationships into structured argument networks.
Overview
Claimify transforms messy discourse (conversations, documents, debates, meeting notes) into analyzable claim structures that reveal:
- Explicit and implicit claims
- Logical relationships (supports/opposes/assumes/contradicts)
- Evidence chains
- Argument structure
- Tension points and gaps
Workflow
- Ingest: Read source material (conversation, document, transcript)
- Extract: Identify atomic claims (one assertion per claim)
- Classify: Label claim types (factual/normative/definitional/causal/predictive)
- Map: Build relationship graph (which claims support/oppose/assume others)
- Analyze: Identify structure, gaps, contradictions, implicit assumptions
- Output: Format as requested (table/graph/narrative/JSON)
Claim Extraction Guidelines
Atomic Claims
Each claim should be a single, testable assertion.
Good:
- "AI adoption increases productivity by 15-30%"
- "Psychological safety enables team learning"
- "Current training methods fail to build AI fluency"
Bad (not atomic):
- "AI is useful and everyone should use it" → Split into 2 claims
Claim Types
| Type | Definition | Example |
|---|---|---|
| Factual | Empirical statement about reality | "Remote work increased 300% since 2020" |
| Normative | Value judgment or prescription | "Organizations should invest in AI training" |
| Definitional | Establishes meaning | "AI fluency = ability to shape context and evaluate output" |
| Causal | X causes Y | "Lack of training causes AI underutilization" |
| Predictive | Future-oriented | "AI adoption will plateau without culture change" |
| Assumption | Unstated premise | [implicit] "Humans resist change" |
Relationship Types
- Supports: Claim A provides evidence/reasoning for claim B
- Opposes: Claim A undermines or contradicts claim B
- Assumes: Claim A requires claim B to be true (often implicit)
- Refines: Claim A specifies/clarifies claim B
- Contradicts: Claims are mutually exclusive
- Independent: No logical relationship
Output Formats
Table Format (default)
| ID | Claim | Type | Supports | Opposes | Assumes | Evidence | |----|-------|------|----------|---------|---------|----------| | C1 | [claim text] | Factual | - | - | C5 | [source/reasoning] | | C2 | [claim text] | Normative | C1 | C4 | - | [source/reasoning] |
Graph Format
Use Mermaid for visualization:
graph TD C1[Claim 1: AI increases productivity] C2[Claim 2: Training is insufficient] C3[Claim 3: Organizations should invest] C1 -->|supports| C3 C2 -->|supports| C3 C2 -.->|assumes| C4[Implicit: Change requires structure]
Narrative Format
Write as structured prose with clear transitions showing logical flow:
## Core Argument The author argues that [main claim]. This rests on three supporting claims: 1. [Factual claim] - This is supported by [evidence] 2. [Causal claim] - However, this assumes [implicit assumption] 3. [Normative claim] - This follows if we accept [prior claims] ## Tensions The argument contains internal tensions: - Claims C2 and C5 appear contradictory because... - The causal chain from C3→C7 has a missing premise...
JSON Format
For programmatic processing:
{ "claims": [ { "id": "C1", "text": "AI adoption increases productivity", "type": "factual", "explicit": true, "supports": ["C3"], "opposed_by": [], "assumes": ["C4"], "evidence": "Multiple case studies cited" } ], "relationships": [ {"from": "C1", "to": "C3", "type": "supports", "strength": "strong"} ], "meta_analysis": { "completeness": "Missing link between C2 and C5", "contradictions": ["C4 vs C7"], "key_assumptions": ["C4", "C9"] } }
Analysis Depth Levels
Level 1: Surface
- Extract only explicit claims
- Basic support/oppose relationships
- No implicit assumption mining
Level 2: Standard (default)
- Extract explicit claims
- Identify clear logical relationships
- Surface obvious implicit assumptions
- Flag apparent contradictions
Level 3: Deep
- Extract all claims (explicit + implicit)
- Map full logical structure
- Identify hidden assumptions
- Analyze argument completeness
- Red-team reasoning
- Suggest strengthening moves
Best Practices
- Be charitable: Steelman arguments before critique
- Distinguish: Separate what's claimed from what's implied
- Be atomic: One claim per line, no compound assertions
- Track evidence: Note source/support for each claim
- Flag uncertainty: Mark inferential leaps
- Mind the gaps: Identify missing premises explicitly
- Stay neutral: Describe structure before evaluating strength
Common Patterns
Argument Chains
Premise 1 (factual) → Premise 2 (causal) → Conclusion (normative)
Implicit Assumptions
Often found by asking: "What must be true for this conclusion to follow?"
Contradictions
Watch for:
- Same speaker, different times
- Different speakers, same topic
- Explicit vs implicit claims
Weak Links
- Unsupported factual claims
- Causal claims without mechanism
- Normative leaps (is → ought)
- Definitional ambiguity
Examples
See
references/examples.md for detailed worked examples.