All Posts
ENApril 4, 2026 5 min read

Testing the Creative Limits of Large Language Models: What CREATE Reveals About Associative Reasoning

The field of AI has made remarkable strides in developing models that can reason, write, and solve complex problems. Yet one frontier remains particularly elusive: genuine creativity. A new benchmark called CREATE, introduced by researchers from New York University and the University of Texas at Austin, provides fresh insights into how well our most advanced language models can engage in associative creativity, the ability to draw novel yet meaningful connections between concepts.

The Challenge of Measuring Machine Creativity

Creativity has long been considered one of the pinnacles of human intelligence, sitting atop Bloom's Taxonomy and forming a core component of Sternberg's theory of intelligence. Yet measuring creativity in artificial systems presents unique challenges. Real-world creative tasks like scientific hypothesis generation are subjective and difficult to evaluate at scale, while abstract symbolic tasks may not reflect how models actually perform on meaningful problems.

The CREATE benchmark attempts to bridge this gap by focusing on associative creativity, specifically the ability to find multiple, diverse pathways connecting concepts through a model's parametric knowledge. Consider the example from the paper: "How is Dakota Johnson connected to people who starred in fantasy/sci-fi movies?" A creative response might note that she starred alongside Chris Evans in "The Materialists," and Evans appeared in "Captain America." But it might also reveal that she's Antonio Banderas' stepdaughter, and Banderas voiced a character in "Shrek."

What makes this approach particularly clever is that it requires models to demonstrate both specificity (the connections must be distinctive and close) and diversity (the paths should be dissimilar from each other). This dual requirement mirrors the core tension in human creativity between quality and novelty.

Methodology and Surprising Findings

The researchers designed CREATE to span different conceptual domains, from entertainment industry connections to genetic associations with diseases. Models are evaluated using a "creative utility" metric that integrates both the quality of individual connections and the diversity across multiple paths. Additionally, a "distinctiveness" metric measures how unique a model's responses are compared to a population of other responses.

The results reveal several counterintuitive findings about current AI capabilities. While frontier models like GPT-4 and Claude do achieve higher creative utility scores than open-source alternatives, they still struggle with generating truly distinctive solutions. More surprisingly, the so-called "thinking models" that use extended reasoning with high token budgets do not consistently outperform on this task.

This finding challenges a fundamental assumption about AI reasoning: that more computational effort necessarily leads to better creative outcomes. The paper's results suggest that chain-of-thought depth and lateral search breadth represent genuinely orthogonal cognitive demands. A model might engage in extensive step-by-step reasoning without ever exploring the diverse conceptual neighborhoods necessary for creative association.

The Depth vs. Breadth Problem in AI Reasoning

The disconnect between reasoning depth and creative breadth reveals something profound about how current language models search conceptual spaces. Traditional chain-of-thought prompting encourages models to follow logical sequences, building conclusions step by step. This approach excels at problems with clear solution paths but may actually constrain the kind of lateral thinking required for creative association.

Consider how humans approach brainstorming tasks. Rather than following a single line of reasoning to its conclusion, creative individuals often engage in rapid conceptual jumps, exploring multiple associative pathways simultaneously. They might start with Dakota Johnson, quickly pivot to her family connections, then leap to voice acting, then to animated films, building a rich network of potential connections before selecting the most interesting ones.

Current language models, despite their impressive capabilities, appear to struggle with this kind of parallel exploration. Their sequential generation process and training objectives may bias them toward coherent, linear reasoning paths rather than the more chaotic, exploratory search patterns that characterize human creativity.

This limitation has broader implications for AI applications in creative domains. Scientific hypothesis generation, artistic creation, and innovative problem-solving all require the ability to traverse unexpected conceptual territories. If our most advanced models cannot effectively balance depth and breadth in their reasoning, they may remain limited partners in genuinely creative endeavors.

Implications for Future AI Development

The CREATE benchmark reveals that creativity in AI systems requires more than just scaling up existing approaches. The finding that extended reasoning tokens do not reliably improve creative performance suggests that we need fundamentally different architectures or training paradigms to support associative thinking.

One promising direction might involve developing models that can maintain multiple competing hypotheses simultaneously, rather than committing to single reasoning paths. Another approach could focus on training objectives that explicitly reward conceptual diversity and unexpected connections, rather than just coherence and accuracy.

The paper also highlights the importance of evaluation methodology in AI research. By creating a benchmark that balances objective evaluation with meaningful creative demands, the researchers provide a valuable tool for measuring progress in this domain. The fact that even frontier models show clear limitations suggests that benchmark saturation is not an immediate concern, giving researchers a stable target for improvement.

Looking forward, the insights from CREATE raise fundamental questions about the nature of machine creativity. Is the sequential, autoregressive generation process inherently limiting for creative tasks? Do we need new architectures that can better support parallel exploration of conceptual spaces? How can we train models to value novelty and diversity without sacrificing quality and coherence?

These questions become increasingly urgent as AI systems are deployed in creative and scientific contexts. Understanding the boundaries of current capabilities, and the cognitive demands that remain challenging, is essential for developing more capable and genuinely creative AI systems.

The CREATE benchmark represents an important step toward more rigorous evaluation of AI creativity, revealing that the path to truly creative machines may require rethinking some of our fundamental assumptions about how artificial intelligence should reason and search through conceptual spaces.

Testing the Creative Limits of Large Language Models: What CREATE Reveals About Associative Reasoning | kualia.ai