The $1 Trillion Question
The $1 Trillion Question, Mortality in the Age of Generative Ghosts, Your Mind and AI, How to Design AI Tutors for Learning, The Imagining Summit Preview: Adam Cutler, and Helen's Book of the Week.
Recent studies claim AI outperforms humans in creativity tests, but these only measure "creative potential." Examples show AI ideas often lack practicality and appeal. The future of AI-enhanced creativity lies in designing tools that allow for exploration, playfulness, and guidance.
Recently, several papers have been published with headlines claiming that AI is more creative than humans. Personally, these headlines make me anxious. I deeply value creativity as an expression of individuality and humanity. My bias is to view our entire human endeavor as centered on creativity—not just in art, music, and traditionally creative professions, but also in how we advance knowledge and solve complex problems.
So, my heart sinks a little each time I read another headline suggesting that GPT-4 surpasses all but the most creative humans. Is AI actually more creative? On key measures of creativity, specifically divergent thinking tasks, it seems so. But we shouldn't be surprised by this finding. Divergent creativity tasks like the Alternative Uses Task (AUT), the Consequences Task (CT), and the Divergent Associations Test (DAT) are fundamentally language-based. Therefore, it's not surprising that large language models excel at these tasks, which measure the semantic distance between words generated in the test.
For example, the AUT asks participants (or AI) to generate alternative uses for common objects, such as a fork. The CT asks for potential consequences of hypothetical scenarios, like "What if humans no longer needed to sleep?" The DAT challenges participants to come up with words that are as different as possible from each other. Creativity in these tasks is measured by fluency, originality, and elaboration—all of which are language-driven metrics.
One of the advantages of AI is its ability to precisely measure these aspects. Semantic distance between words or concepts is no longer solely a human judgment. Language models inherently run on this principle, making related metrics much easier to obtain. This is similar to any measurement bias in that once we can measure something accurately, we start to uncover more about it.
The Artificiality Weekend Briefing: About AI, Not Written by AI