creative-thinking-for-research
Creative Thinking for Research
Eight empirically grounded frameworks from cognitive science, applied to computer science and AI research. Unlike ad-hoc brainstorming, each framework here is backed by decades of creativity research — from Koestler's bisociation to Kauffman's adjacent possible. They target distinct cognitive operations: combining, reformulating, analogizing, constraining, inverting, abstracting, exploring boundaries, and holding contradictions.
When to Use This Skill
- Generating genuinely novel ideas, not incremental extensions of prior work
- Feeling trapped in a local optimum of thinking within a single subfield
- Wanting to systematically apply creativity heuristics rather than waiting for inspiration
- Preparing for a research retreat or PhD-level ideation session
- Bridging between fields and seeking structural (not superficial) connections
Do NOT use this skill when:
- You need structured project-level brainstorming workflows (use
brainstorming-research-ideas) - You have a well-defined problem and need execution help (use domain-specific skills)
- You need a literature survey (use
scientific-skills:literature-review)
Relationship to Brainstorm skill: The brainstorm skill provides operational workflows (diverge → converge → refine) and practical filters. This skill provides the deeper cognitive engines that power creative leaps. Use them together: creative-thinking to generate raw insight, brainstorm to structure and evaluate it.
Framework 1: Combinatorial Creativity (Bisociation)
Novel ideas arise from combining existing concepts in unexpected ways. Arthur Koestler called this bisociation — connecting two previously unrelated frames of reference, as distinct from routine association within a single frame.
Why it works: Meta-research consistently shows that breadth of knowledge is a precursor to creative output. People who read across disciplines produce more novel work. The combination itself is the creative act.
In CS Research:
- Biological evolution → optimization (genetic algorithms)
- Game theory → networking (mechanism design for routing)
- Statistical physics → machine learning (Boltzmann machines, energy-based models)
- Linguistics → programming (type theory, formal grammars)
Systematic Bisociation Workflow:
- Select two domains you have at least passing familiarity with
- List core primitives in each domain (5-10 fundamental concepts per domain)
- Create a cross-product matrix: row = concepts from Domain A, column = concepts from Domain B
- For each cell, ask: "What would it mean to apply A's concept to B's problem?"
- Filter: Which combinations produce a non-trivial, testable research question?
- Validate structural depth: Is the connection mechanistic or merely metaphorical?
Cross-Product Example:
| Caching | Load Balancing | Fault Tolerance | |
|---|---|---|---|
| Natural Selection | Evict least-fit entries | Adaptive allocation via fitness | Population-level redundancy |
| Immune Memory | Learned threat signatures | Distributed detection | Self/non-self discrimination |
| Symbiosis | Cooperative prefetching | Mutualistic resource sharing | Co-dependent resilience |
Quality Test: A strong bisociation is not a surface metaphor ("the network is like a brain") but a structural mapping where the mechanism transfers ("attention mechanisms implement a form of selective gating analogous to cognitive attention filtering").
Self-Check:
- Is the connection structural (mechanisms map) or merely verbal (labels map)?
- Does the combination generate testable predictions?
- Would an expert in both fields find the connection non-obvious but sound?
Framework 2: Problem Reformulation (Representational Change)
Gestalt psychologists identified that breakthroughs often come not from solving the problem as stated, but from re-representing the problem itself. Kaplan and Simon's work on insight shows that changing the problem space — the constraints, the abstraction level, the formalism — is often where creativity lives.
The Key Shift: From "How do I solve this problem?" to "Am I even thinking about this problem correctly?"
Reformulation Strategies:
| Strategy | Example |
|---|---|
| Change the objective | "Make the algorithm faster" → "Eliminate the need for this computation" |
| Change the formalism | Graph problem → linear algebra problem (spectral methods) |
| Change the granularity | Per-token prediction → per-span prediction |
| Change the agent | "How should the model learn?" → "How should the data teach?" (curriculum learning) |
| Change the timescale | Real-time optimization → amortized inference |
| Invert the direction | Forward simulation → inverse problem (learning from observations) |
Workflow:
- State your current problem in one sentence
- Identify the hidden assumptions in that statement:
- What formalism are you using? (Could you use a different one?)
- What is the objective? (Is it the right objective?)
- What level of granularity? (Could you go coarser or finer?)
- Who is the agent? (Could you shift perspective?)
- For each assumption, generate the alternative: "What if [opposite assumption]?"
- For each alternative, ask: "Does this reformulation make the problem easier, harder, or different in a useful way?"
- A reformulation that makes a hard problem easy is often a publishable insight on its own
Classic CS Examples:
- PageRank: Reformulated "find important web pages" from content analysis to graph eigenvalue problem
- Dropout: Reformulated "prevent overfitting" from regularization to approximate ensemble
- Attention: Reformulated "handle long sequences" from remembering everything to selectively querying
Framework 3: Analogical Reasoning (Structure-Mapping)
Dedre Gentner's structure-mapping theory and Kevin Dunbar's studies of real scientists show that analogy is the core engine of scientific creativity. The critical finding: surface-level analogies are common but weak; structural or relational analogies — where the deep causal/relational structure maps across domains — produce the most powerful insights.
Dunbar's Finding: In the most successful labs, analogies from distant domains drove the most important discoveries. Nearby analogies refined ideas; distant analogies generated them.
Levels of Analogical Depth:
| Level | Description | Value | Example |
|---|---|---|---|
| Surface | Things look similar | Low | "A neural network is like a brain" |
| Relational | Relationships between entities match | Medium | "Attention allocation in models parallels resource allocation in economics" |
| Structural | Deep causal mechanisms map | High | "Diffusion models reverse a thermodynamic process; the math of non-equilibrium stat-mech directly applies" |
Structure-Mapping Workflow:
- Describe your problem using only relational/causal language (strip domain-specific nouns)
- Bad: "We need to improve transformer attention efficiency"
- Good: "We have a system that must selectively aggregate information from a large set, where relevance is context-dependent and the cost scales quadratically with set size"
- Search for structural matches: What other systems selectively aggregate from large sets?
- Database query optimization, visual attention in neuroscience, information retrieval, resource allocation
- Pick the most distant match with genuine structural fidelity
- Map the solution mechanism: How does the source domain solve this?
- Transfer and adapt: What changes when you bring that mechanism into your domain?
- Generate predictions: The analogy should tell you something you didn't already know
Validation Checklist:
- Does the mapping preserve causal/relational structure (not just labels)?
- Can I identify at least one prediction the analogy makes in my domain?
- Would an expert in the source domain confirm the mechanism is correctly understood?
- Is the analogy non-obvious to my target audience?
Framework 4: Constraint Manipulation (Boden's Framework)
Margaret Boden's framework distinguishes three forms of creativity based on how they interact with constraints:
| Type | Operation | CS Example |
|---|---|---|
| Exploratory | Search within the existing conceptual space | Hyperparameter tuning, architecture search within a fixed paradigm |
| Combinational | Combine elements from different spaces | Multi-task learning, neuro-symbolic methods |
| Transformational | Change the rules of the space itself | Dropping the assumption that training requires labels (self-supervised learning) |
Transformational creativity is the rarest and highest-impact. It happens when you change what is even considered a valid solution.
Constraint Analysis Workflow:
- List the constraints of your current approach (5-10 constraints):
- Computational: "Must fit in GPU memory"
- Methodological: "Requires labeled data"
- Architectural: "Uses fixed-length context"
- Evaluative: "Measured by accuracy on benchmark X"
- Classify each constraint:
- Hard: Physically or logically necessary (cannot violate)
- Soft: Convention or historical accident (can question)
- Hidden: Not stated but implicitly assumed (most fertile for innovation)
- For each soft/hidden constraint, ask:
- What if we relaxed it? (streaming algorithms from relaxing "fits in memory")
- What if we tightened it? (efficiency research from tightening compute budgets)
- What if we replaced it with a different constraint entirely?
- The most productive move is often exposing and dropping a hidden constraint
Classic Examples of Constraint Transformation:
- "Data must fit in memory" → dropped → streaming algorithms, external memory
- "Training requires human labels" → dropped → self-supervised learning
- "Models must be deterministic" → dropped → variational methods, diffusion
- "Inference must happen in one pass" → dropped → iterative refinement, chain-of-thought
Framework 5: Negation and Inversion
Take a core assumption in your field and negate it. This is formalized in De Bono's lateral thinking and the TRIZ methodology from engineering.
The Pattern: "What if [widely held assumption] is wrong, unnecessary, or invertible?"
Systematic Negation Workflow:
- List 5-10 core assumptions in your subfield (the things "everyone knows")
- Negate each one and ask: What system would you build?
- Evaluate each negation:
- Incoherent → discard
- Already explored → check if conditions have changed (see brainstorm skill, Framework 5)
- Unexplored and coherent → potential research direction
Negation Hall of Fame in CS:
| Assumption | Negation | Result |
|---|---|---|
| "We need strong consistency" | What if we don't? | Eventual consistency, CRDTs |
| "We need exact answers" | What if approximate is fine? | Sketches, LSH, approximate nearest neighbors |
| "Labels are necessary" | What if we learn without them? | Self-supervised learning, contrastive methods |
| "More parameters = more compute" | What if we don't use all parameters? | Mixture of Experts, sparse models |
| "Training and inference are separate" | What if the model keeps learning? | Online learning, test-time training |
| "Errors must be prevented" | What if we embrace and correct them? | Speculative decoding, self-correction |
TRIZ-Inspired Principles for CS:
| TRIZ Principle | CS Application |
|---|---|
| Inversion | Reverse the process (generative vs. discriminative) |
| Segmentation | Break monolithic into modular (microservices, mixture of experts) |
| Merging | Combine separate steps (end-to-end learning) |
| Universality | One component serves multiple functions (multi-task models) |
| Nesting | Place one system inside another (meta-learning) |
| Dynamization | Make static things adaptive (dynamic architectures, adaptive computation) |
Framework 6: Abstraction and Generalization Laddering
Moving up and down the abstraction ladder is a fundamental creative act. Polya's heuristics formalize this: "Can you solve a more general problem? A more specific one? An analogous one?"
Three Moves:
| Move | Question | Outcome |
|---|---|---|
| Generalize | "Is my solution a special case of something broader?" | Framework papers, unifying theories |
| Specialize | "What happens when I add extreme constraints?" | Niche applications, surprising edge cases |
| Analogize | "Where else does this abstract pattern appear?" | Cross-domain transfer (see Framework 3) |
Generalization Workflow:
- State your specific result
- Replace each specific element with a variable: "ResNet works for ImageNet" → "Architecture X works for distribution Y"
- Ask: Under what conditions does this hold? What is the general principle?
- If the general principle is novel → that is the contribution
Specialization Workflow:
- Take a general method
- Add extreme constraints: tiny data, huge dimensionality, adversarial inputs, real-time requirements
- Ask: Does the method still work? If not, why not?
- The failure case often reveals the method's true assumptions
When to Generalize vs. Specialize:
- Generalize when you have results but no explanation
- Specialize when you have theory but no grounding
- Analogize when you are stuck in either direction
Framework 7: The Adjacent Possible (Kauffman / Johnson)
Stuart Kauffman's concept, popularized by Steven Johnson: innovation happens at the boundary of what is currently reachable — the adjacent possible. New ideas become thinkable once their prerequisites exist. This explains why simultaneous independent discovery is so common — multiple people reach the same boundary.
Practical Implication: Map what has recently become possible and explore the space those enablers open.
Adjacent Possible Mapping Workflow:
- List recent enablers (last 1-3 years):
- New hardware capabilities (longer context, faster inference, new accelerators)
- New datasets or benchmarks
- New open-source tools or frameworks
- New theoretical results
- New regulatory or social conditions
- For each enabler, ask: "What was previously impossible or impractical that this now permits?"
- Combine enablers: The most powerful adjacent possibles arise from the intersection of multiple new enablers
- Check for competition: If many people can see the same adjacent possible, speed or a unique angle matters
Current Adjacent Possibles (2025-2026):
| Enabler | Newly Possible |
|---|---|
| 1M+ token context windows | Full-codebase reasoning, book-length analysis |
| Inference cost drops (100x in 2 years) | Real-time agentic loops, always-on AI assistants |
| Open-weight models at GPT-4 level | Reproducible research on frontier capabilities |
| Multimodal models (vision + language + audio) | Unified perception-reasoning systems |
| Synthetic data at scale | Training data for domains with no natural data |
| Tool-using models | Research automation, self-improving systems |
Timing Signal: If your idea requires technology that doesn't exist yet, it's beyond the adjacent possible — park it. If your idea could have been done 5 years ago, someone probably did — check the literature. The sweet spot is ideas that became feasible in the last 6-18 months.
Framework 8: Janusian and Dialectical Thinking
Albert Rothenberg's studies of eminent creators found that holding two contradictory ideas simultaneously is a hallmark of creative thinking. Named after Janus, the two-faced Roman god, this mode of thinking doesn't resolve contradictions by choosing a side — it generates new frameworks that transcend the opposition.
In CS: The most influential results often emerge from tensions previously thought irreconcilable.
| Contradiction | Resolution | Impact |
|---|---|---|
| Consistency AND Availability (distributed systems) | CAP theorem: formalized the trade-off, then Raft/CRDTs found practical middle grounds | Foundation of distributed systems theory |
| Security AND Usability | Zero-knowledge proofs: prove knowledge without revealing it | Enabled private computation |
| Expressiveness AND Tractability | Probabilistic programming: express complex models, automate inference | New programming paradigm |
| Memorization AND Generalization | Grokking: models memorize first, then generalize with more training | New understanding of learning dynamics |
| Compression AND Quality | Neural codecs that compress beyond information-theoretic limits via learned priors | Redefined compression research |
Dialectical Thinking Workflow:
- Identify a binary in your field: A vs. B (two approaches, goals, or paradigms treated as opposites)
- Resist choosing a side. Instead ask:
- "What would a system look like that achieves both A and B?"
- "Under what conditions is the A-B trade-off not fundamental?"
- "Is the opposition an artifact of how we formalized the problem?"
- Seek synthesis: The resolution often requires a new abstraction that reframes the relationship
- Test the synthesis: Can you demonstrate empirically that both goals are achievable?
Self-Check:
- Am I holding the contradiction genuinely (not prematurely resolving it)?
- Is the synthesis a new idea, not just a compromise (splitting the difference)?
- Does the resolution change how people think about the problem, not just the solution?
Combining Frameworks: A Creative Thinking Protocol
These frameworks are most powerful in combination. Here is a systematic protocol for a deep creative thinking session:
Phase 1: Map the Space (15 min)
- Constraint Manipulation (F4): List all constraints of the current paradigm. Mark which are hard, soft, hidden.
- Adjacent Possible (F7): List recent enablers that change the feasibility landscape.
Phase 2: Generate Disruptions (30 min)
- Negation (F5): Negate 3 soft/hidden constraints. What systems emerge?
- Bisociation (F1): Pick a distant field and create a cross-product matrix with your domain.
- Problem Reformulation (F2): Restate your problem 3 different ways (change objective, formalism, agent).
Phase 3: Deepen Promising Leads (30 min)
- Analogical Reasoning (F3): For each promising idea, find a structural analogy and extract predictions.
- Abstraction Laddering (F6): Move each idea up (generalize) and down (specialize).
- Janusian Thinking (F8): Identify any tensions. Can you synthesize rather than choose?
Phase 4: Evaluate (15 min)
Apply the two-sentence test (from the brainstorm skill):
"[Domain] currently struggles with [problem] because [reason]. We [approach] by [mechanism], which works because [insight]."
Any idea that survives all four phases and passes the two-sentence test is worth pursuing.
Common Creative Blocks and Unblocking Strategies
| Block | Symptom | Framework to Apply |
|---|---|---|
| Fixation | Cannot stop thinking about the problem one way | Problem Reformulation (F2) — force a different representation |
| Tunnel vision | All ideas come from the same subfield | Bisociation (F1) or Analogical Reasoning (F3) — import from elsewhere |
| Self-censoring | Dismissing ideas as "too weird" before exploring | Negation (F5) — weird is the point; evaluate after generating |
| Incrementalism | Every idea is "+2% on benchmark X" | Constraint Manipulation (F4) — change the rules, not the parameters |
| Analysis paralysis | Too many options, cannot commit | Adjacent Possible (F7) — what is feasible right now? |
| False dichotomy | Stuck choosing between two approaches | Janusian Thinking (F8) — seek synthesis, not selection |
Usage Instructions for Agents
When a researcher asks for help with creative thinking or novel ideation:
- Assess the block: What kind of thinking are they stuck in? (See Common Creative Blocks table)
- Select 2-3 frameworks based on the block type
- Walk through each framework interactively, asking the researcher to supply domain-specific content
- Push for structural depth: If an analogy or combination is surface-level, probe deeper
- Maintain a running list of all generated ideas, even unusual ones
- Apply the two-sentence test to candidates that survive exploration
- Hand off to the brainstorm skill for systematic evaluation (diverge → converge → refine)
Key Principles:
- Generative mode first, evaluative mode second — do not filter prematurely
- Distant analogies are more valuable than nearby ones, but require more validation
- The researcher's domain expertise is essential — the agent provides the cognitive scaffolding, not the domain knowledge
- Encourage the researcher to sit with contradictions rather than resolve them quickly