Skillshub creative-thinking-for-research

Creative Thinking for Research

install
source · Clone the upstream repo
git clone https://github.com/ComeOnOliver/skillshub
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/ComeOnOliver/skillshub "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/Orchestra-Research/AI-Research-SKILLs/creative-thinking-for-research" ~/.claude/skills/comeonoliver-skillshub-creative-thinking-for-research && rm -rf "$T"
manifest: skills/Orchestra-Research/AI-Research-SKILLs/creative-thinking-for-research/SKILL.md
source content

Creative Thinking for Research

Eight empirically grounded frameworks from cognitive science, applied to computer science and AI research. Unlike ad-hoc brainstorming, each framework here is backed by decades of creativity research — from Koestler's bisociation to Kauffman's adjacent possible. They target distinct cognitive operations: combining, reformulating, analogizing, constraining, inverting, abstracting, exploring boundaries, and holding contradictions.

When to Use This Skill

  • Generating genuinely novel ideas, not incremental extensions of prior work
  • Feeling trapped in a local optimum of thinking within a single subfield
  • Wanting to systematically apply creativity heuristics rather than waiting for inspiration
  • Preparing for a research retreat or PhD-level ideation session
  • Bridging between fields and seeking structural (not superficial) connections

Do NOT use this skill when:

  • You need structured project-level brainstorming workflows (use
    brainstorming-research-ideas
    )
  • You have a well-defined problem and need execution help (use domain-specific skills)
  • You need a literature survey (use
    scientific-skills:literature-review
    )

Relationship to Brainstorm skill: The brainstorm skill provides operational workflows (diverge → converge → refine) and practical filters. This skill provides the deeper cognitive engines that power creative leaps. Use them together: creative-thinking to generate raw insight, brainstorm to structure and evaluate it.


Framework 1: Combinatorial Creativity (Bisociation)

Novel ideas arise from combining existing concepts in unexpected ways. Arthur Koestler called this bisociation — connecting two previously unrelated frames of reference, as distinct from routine association within a single frame.

Why it works: Meta-research consistently shows that breadth of knowledge is a precursor to creative output. People who read across disciplines produce more novel work. The combination itself is the creative act.

In CS Research:

  • Biological evolution → optimization (genetic algorithms)
  • Game theory → networking (mechanism design for routing)
  • Statistical physics → machine learning (Boltzmann machines, energy-based models)
  • Linguistics → programming (type theory, formal grammars)

Systematic Bisociation Workflow:

  1. Select two domains you have at least passing familiarity with
  2. List core primitives in each domain (5-10 fundamental concepts per domain)
  3. Create a cross-product matrix: row = concepts from Domain A, column = concepts from Domain B
  4. For each cell, ask: "What would it mean to apply A's concept to B's problem?"
  5. Filter: Which combinations produce a non-trivial, testable research question?
  6. Validate structural depth: Is the connection mechanistic or merely metaphorical?

Cross-Product Example:

CachingLoad BalancingFault Tolerance
Natural SelectionEvict least-fit entriesAdaptive allocation via fitnessPopulation-level redundancy
Immune MemoryLearned threat signaturesDistributed detectionSelf/non-self discrimination
SymbiosisCooperative prefetchingMutualistic resource sharingCo-dependent resilience

Quality Test: A strong bisociation is not a surface metaphor ("the network is like a brain") but a structural mapping where the mechanism transfers ("attention mechanisms implement a form of selective gating analogous to cognitive attention filtering").

Self-Check:

  • Is the connection structural (mechanisms map) or merely verbal (labels map)?
  • Does the combination generate testable predictions?
  • Would an expert in both fields find the connection non-obvious but sound?

Framework 2: Problem Reformulation (Representational Change)

Gestalt psychologists identified that breakthroughs often come not from solving the problem as stated, but from re-representing the problem itself. Kaplan and Simon's work on insight shows that changing the problem space — the constraints, the abstraction level, the formalism — is often where creativity lives.

The Key Shift: From "How do I solve this problem?" to "Am I even thinking about this problem correctly?"

Reformulation Strategies:

StrategyExample
Change the objective"Make the algorithm faster" → "Eliminate the need for this computation"
Change the formalismGraph problem → linear algebra problem (spectral methods)
Change the granularityPer-token prediction → per-span prediction
Change the agent"How should the model learn?" → "How should the data teach?" (curriculum learning)
Change the timescaleReal-time optimization → amortized inference
Invert the directionForward simulation → inverse problem (learning from observations)

Workflow:

  1. State your current problem in one sentence
  2. Identify the hidden assumptions in that statement:
    • What formalism are you using? (Could you use a different one?)
    • What is the objective? (Is it the right objective?)
    • What level of granularity? (Could you go coarser or finer?)
    • Who is the agent? (Could you shift perspective?)
  3. For each assumption, generate the alternative: "What if [opposite assumption]?"
  4. For each alternative, ask: "Does this reformulation make the problem easier, harder, or different in a useful way?"
  5. A reformulation that makes a hard problem easy is often a publishable insight on its own

Classic CS Examples:

  • PageRank: Reformulated "find important web pages" from content analysis to graph eigenvalue problem
  • Dropout: Reformulated "prevent overfitting" from regularization to approximate ensemble
  • Attention: Reformulated "handle long sequences" from remembering everything to selectively querying

Framework 3: Analogical Reasoning (Structure-Mapping)

Dedre Gentner's structure-mapping theory and Kevin Dunbar's studies of real scientists show that analogy is the core engine of scientific creativity. The critical finding: surface-level analogies are common but weak; structural or relational analogies — where the deep causal/relational structure maps across domains — produce the most powerful insights.

Dunbar's Finding: In the most successful labs, analogies from distant domains drove the most important discoveries. Nearby analogies refined ideas; distant analogies generated them.

Levels of Analogical Depth:

LevelDescriptionValueExample
SurfaceThings look similarLow"A neural network is like a brain"
RelationalRelationships between entities matchMedium"Attention allocation in models parallels resource allocation in economics"
StructuralDeep causal mechanisms mapHigh"Diffusion models reverse a thermodynamic process; the math of non-equilibrium stat-mech directly applies"

Structure-Mapping Workflow:

  1. Describe your problem using only relational/causal language (strip domain-specific nouns)
    • Bad: "We need to improve transformer attention efficiency"
    • Good: "We have a system that must selectively aggregate information from a large set, where relevance is context-dependent and the cost scales quadratically with set size"
  2. Search for structural matches: What other systems selectively aggregate from large sets?
    • Database query optimization, visual attention in neuroscience, information retrieval, resource allocation
  3. Pick the most distant match with genuine structural fidelity
  4. Map the solution mechanism: How does the source domain solve this?
  5. Transfer and adapt: What changes when you bring that mechanism into your domain?
  6. Generate predictions: The analogy should tell you something you didn't already know

Validation Checklist:

  • Does the mapping preserve causal/relational structure (not just labels)?
  • Can I identify at least one prediction the analogy makes in my domain?
  • Would an expert in the source domain confirm the mechanism is correctly understood?
  • Is the analogy non-obvious to my target audience?

Framework 4: Constraint Manipulation (Boden's Framework)

Margaret Boden's framework distinguishes three forms of creativity based on how they interact with constraints:

TypeOperationCS Example
ExploratorySearch within the existing conceptual spaceHyperparameter tuning, architecture search within a fixed paradigm
CombinationalCombine elements from different spacesMulti-task learning, neuro-symbolic methods
TransformationalChange the rules of the space itselfDropping the assumption that training requires labels (self-supervised learning)

Transformational creativity is the rarest and highest-impact. It happens when you change what is even considered a valid solution.

Constraint Analysis Workflow:

  1. List the constraints of your current approach (5-10 constraints):
    • Computational: "Must fit in GPU memory"
    • Methodological: "Requires labeled data"
    • Architectural: "Uses fixed-length context"
    • Evaluative: "Measured by accuracy on benchmark X"
  2. Classify each constraint:
    • Hard: Physically or logically necessary (cannot violate)
    • Soft: Convention or historical accident (can question)
    • Hidden: Not stated but implicitly assumed (most fertile for innovation)
  3. For each soft/hidden constraint, ask:
    • What if we relaxed it? (streaming algorithms from relaxing "fits in memory")
    • What if we tightened it? (efficiency research from tightening compute budgets)
    • What if we replaced it with a different constraint entirely?
  4. The most productive move is often exposing and dropping a hidden constraint

Classic Examples of Constraint Transformation:

  • "Data must fit in memory" → dropped → streaming algorithms, external memory
  • "Training requires human labels" → dropped → self-supervised learning
  • "Models must be deterministic" → dropped → variational methods, diffusion
  • "Inference must happen in one pass" → dropped → iterative refinement, chain-of-thought

Framework 5: Negation and Inversion

Take a core assumption in your field and negate it. This is formalized in De Bono's lateral thinking and the TRIZ methodology from engineering.

The Pattern: "What if [widely held assumption] is wrong, unnecessary, or invertible?"

Systematic Negation Workflow:

  1. List 5-10 core assumptions in your subfield (the things "everyone knows")
  2. Negate each one and ask: What system would you build?
  3. Evaluate each negation:
    • Incoherent → discard
    • Already explored → check if conditions have changed (see brainstorm skill, Framework 5)
    • Unexplored and coherent → potential research direction

Negation Hall of Fame in CS:

AssumptionNegationResult
"We need strong consistency"What if we don't?Eventual consistency, CRDTs
"We need exact answers"What if approximate is fine?Sketches, LSH, approximate nearest neighbors
"Labels are necessary"What if we learn without them?Self-supervised learning, contrastive methods
"More parameters = more compute"What if we don't use all parameters?Mixture of Experts, sparse models
"Training and inference are separate"What if the model keeps learning?Online learning, test-time training
"Errors must be prevented"What if we embrace and correct them?Speculative decoding, self-correction

TRIZ-Inspired Principles for CS:

TRIZ PrincipleCS Application
InversionReverse the process (generative vs. discriminative)
SegmentationBreak monolithic into modular (microservices, mixture of experts)
MergingCombine separate steps (end-to-end learning)
UniversalityOne component serves multiple functions (multi-task models)
NestingPlace one system inside another (meta-learning)
DynamizationMake static things adaptive (dynamic architectures, adaptive computation)

Framework 6: Abstraction and Generalization Laddering

Moving up and down the abstraction ladder is a fundamental creative act. Polya's heuristics formalize this: "Can you solve a more general problem? A more specific one? An analogous one?"

Three Moves:

MoveQuestionOutcome
Generalize"Is my solution a special case of something broader?"Framework papers, unifying theories
Specialize"What happens when I add extreme constraints?"Niche applications, surprising edge cases
Analogize"Where else does this abstract pattern appear?"Cross-domain transfer (see Framework 3)

Generalization Workflow:

  1. State your specific result
  2. Replace each specific element with a variable: "ResNet works for ImageNet" → "Architecture X works for distribution Y"
  3. Ask: Under what conditions does this hold? What is the general principle?
  4. If the general principle is novel → that is the contribution

Specialization Workflow:

  1. Take a general method
  2. Add extreme constraints: tiny data, huge dimensionality, adversarial inputs, real-time requirements
  3. Ask: Does the method still work? If not, why not?
  4. The failure case often reveals the method's true assumptions

When to Generalize vs. Specialize:

  • Generalize when you have results but no explanation
  • Specialize when you have theory but no grounding
  • Analogize when you are stuck in either direction

Framework 7: The Adjacent Possible (Kauffman / Johnson)

Stuart Kauffman's concept, popularized by Steven Johnson: innovation happens at the boundary of what is currently reachable — the adjacent possible. New ideas become thinkable once their prerequisites exist. This explains why simultaneous independent discovery is so common — multiple people reach the same boundary.

Practical Implication: Map what has recently become possible and explore the space those enablers open.

Adjacent Possible Mapping Workflow:

  1. List recent enablers (last 1-3 years):
    • New hardware capabilities (longer context, faster inference, new accelerators)
    • New datasets or benchmarks
    • New open-source tools or frameworks
    • New theoretical results
    • New regulatory or social conditions
  2. For each enabler, ask: "What was previously impossible or impractical that this now permits?"
  3. Combine enablers: The most powerful adjacent possibles arise from the intersection of multiple new enablers
  4. Check for competition: If many people can see the same adjacent possible, speed or a unique angle matters

Current Adjacent Possibles (2025-2026):

EnablerNewly Possible
1M+ token context windowsFull-codebase reasoning, book-length analysis
Inference cost drops (100x in 2 years)Real-time agentic loops, always-on AI assistants
Open-weight models at GPT-4 levelReproducible research on frontier capabilities
Multimodal models (vision + language + audio)Unified perception-reasoning systems
Synthetic data at scaleTraining data for domains with no natural data
Tool-using modelsResearch automation, self-improving systems

Timing Signal: If your idea requires technology that doesn't exist yet, it's beyond the adjacent possible — park it. If your idea could have been done 5 years ago, someone probably did — check the literature. The sweet spot is ideas that became feasible in the last 6-18 months.


Framework 8: Janusian and Dialectical Thinking

Albert Rothenberg's studies of eminent creators found that holding two contradictory ideas simultaneously is a hallmark of creative thinking. Named after Janus, the two-faced Roman god, this mode of thinking doesn't resolve contradictions by choosing a side — it generates new frameworks that transcend the opposition.

In CS: The most influential results often emerge from tensions previously thought irreconcilable.

ContradictionResolutionImpact
Consistency AND Availability (distributed systems)CAP theorem: formalized the trade-off, then Raft/CRDTs found practical middle groundsFoundation of distributed systems theory
Security AND UsabilityZero-knowledge proofs: prove knowledge without revealing itEnabled private computation
Expressiveness AND TractabilityProbabilistic programming: express complex models, automate inferenceNew programming paradigm
Memorization AND GeneralizationGrokking: models memorize first, then generalize with more trainingNew understanding of learning dynamics
Compression AND QualityNeural codecs that compress beyond information-theoretic limits via learned priorsRedefined compression research

Dialectical Thinking Workflow:

  1. Identify a binary in your field: A vs. B (two approaches, goals, or paradigms treated as opposites)
  2. Resist choosing a side. Instead ask:
    • "What would a system look like that achieves both A and B?"
    • "Under what conditions is the A-B trade-off not fundamental?"
    • "Is the opposition an artifact of how we formalized the problem?"
  3. Seek synthesis: The resolution often requires a new abstraction that reframes the relationship
  4. Test the synthesis: Can you demonstrate empirically that both goals are achievable?

Self-Check:

  • Am I holding the contradiction genuinely (not prematurely resolving it)?
  • Is the synthesis a new idea, not just a compromise (splitting the difference)?
  • Does the resolution change how people think about the problem, not just the solution?

Combining Frameworks: A Creative Thinking Protocol

These frameworks are most powerful in combination. Here is a systematic protocol for a deep creative thinking session:

Phase 1: Map the Space (15 min)

  1. Constraint Manipulation (F4): List all constraints of the current paradigm. Mark which are hard, soft, hidden.
  2. Adjacent Possible (F7): List recent enablers that change the feasibility landscape.

Phase 2: Generate Disruptions (30 min)

  1. Negation (F5): Negate 3 soft/hidden constraints. What systems emerge?
  2. Bisociation (F1): Pick a distant field and create a cross-product matrix with your domain.
  3. Problem Reformulation (F2): Restate your problem 3 different ways (change objective, formalism, agent).

Phase 3: Deepen Promising Leads (30 min)

  1. Analogical Reasoning (F3): For each promising idea, find a structural analogy and extract predictions.
  2. Abstraction Laddering (F6): Move each idea up (generalize) and down (specialize).
  3. Janusian Thinking (F8): Identify any tensions. Can you synthesize rather than choose?

Phase 4: Evaluate (15 min)

Apply the two-sentence test (from the brainstorm skill):

"[Domain] currently struggles with [problem] because [reason]. We [approach] by [mechanism], which works because [insight]."

Any idea that survives all four phases and passes the two-sentence test is worth pursuing.


Common Creative Blocks and Unblocking Strategies

BlockSymptomFramework to Apply
FixationCannot stop thinking about the problem one wayProblem Reformulation (F2) — force a different representation
Tunnel visionAll ideas come from the same subfieldBisociation (F1) or Analogical Reasoning (F3) — import from elsewhere
Self-censoringDismissing ideas as "too weird" before exploringNegation (F5) — weird is the point; evaluate after generating
IncrementalismEvery idea is "+2% on benchmark X"Constraint Manipulation (F4) — change the rules, not the parameters
Analysis paralysisToo many options, cannot commitAdjacent Possible (F7) — what is feasible right now?
False dichotomyStuck choosing between two approachesJanusian Thinking (F8) — seek synthesis, not selection

Usage Instructions for Agents

When a researcher asks for help with creative thinking or novel ideation:

  1. Assess the block: What kind of thinking are they stuck in? (See Common Creative Blocks table)
  2. Select 2-3 frameworks based on the block type
  3. Walk through each framework interactively, asking the researcher to supply domain-specific content
  4. Push for structural depth: If an analogy or combination is surface-level, probe deeper
  5. Maintain a running list of all generated ideas, even unusual ones
  6. Apply the two-sentence test to candidates that survive exploration
  7. Hand off to the brainstorm skill for systematic evaluation (diverge → converge → refine)

Key Principles:

  • Generative mode first, evaluative mode second — do not filter prematurely
  • Distant analogies are more valuable than nearby ones, but require more validation
  • The researcher's domain expertise is essential — the agent provides the cognitive scaffolding, not the domain knowledge
  • Encourage the researcher to sit with contradictions rather than resolve them quickly