Technology

Researchers uncover hidden ingredients behind AI creativity

By Webb Wright

Copyright livescience

Researchers uncover hidden ingredients behind AI creativity

Skip to main content

Close main menu

Live Science

Sign up to our newsletter

View Profile

Search Live Science

Planet Earth

Archaeology

Physics & Math

Human Behavior

Science news

Life’s Little Mysteries

Science quizzes

Newsletters

Story archive

Man regains vision by tooth to eye implant
5,000-year-old stone tomb in Spain
Abandoning daylight saving time
JWST discovers possible black hole star
Million-year-old skull from China rewrites human origins

Don’t miss these

Artificial Intelligence
Would outsourcing everything to AI cost us our ability to think for ourselves?

Artificial Intelligence
Why OpenAI’s solution to AI hallucinations would kill ChatGPT tomorrow

Artificial Intelligence
AI can’t solve these puzzles that take humans only seconds

Artificial Intelligence
Scientists asked ChatGPT to solve a math problem from more than 2,000 years ago — how it answered it surprised them

Artificial Intelligence
AI could use online images as a backdoor into your computer, alarming new study suggests

Artificial Intelligence
AI outsmarted 30 of the world’s top mathematicians at secret meeting in California

Artificial Intelligence
AI could soon think in ways we don’t even understand — evading our efforts to keep it aligned — top AI scientists warn

Artificial Intelligence
AI models can send subliminal messages that teach other AIs to be ‘evil,’ study claims

Artificial Intelligence
The more advanced AI models get, the better they are at deceiving us — they even know when they’re being tested

This obscure, 80-year-old machine might be the key to unlocking the full potential of AI today

Artificial Intelligence
AI chatbots oversimplify scientific studies and gloss over critical details — the newest models are especially guilty

Artificial Intelligence
OpenAI’s ChatGPT agent can control your PC to do tasks on your behalf — but how does it work and what’s the point?

Artificial Intelligence
New AI system can ‘predict human behavior in any situation’ with unprecedented degree of accuracy, scientists say

Quantum Physics
‘Paraparticles’ would be a third kingdom of quantum particle

Artificial Intelligence
There are 32 different ways AI can go rogue, scientists say — from hallucinating answers to a complete misalignment with humanity

Artificial Intelligence

Researchers uncover hidden ingredients behind AI creativity

Webb Wright

27 September 2025

Image generators are designed to mimic their training data, so where does their apparent creativity come from? A recent study suggests that it’s an inevitable by-product of their architecture.

When you purchase through links on our site, we may earn an affiliate commission. Here’s how it works.

(Image credit: Adrián Astorgano for Quanta Magazine)

We were once promised self-driving cars and robot maids. Instead, we’ve seen the rise of artificial intelligence systems that can beat us in chess, analyze huge reams of text and compose sonnets. This has been one of the great surprises of the modern era: physical tasks that are easy for humans turn out to be very difficult for robots, while algorithms are increasingly able to mimic our intellect.

Another surprise that has long perplexed researchers is those algorithms’ knack for their own, strange kind of creativity.
Diffusion models, the backbone of image-generating tools such as DALL·E, Imagen and Stable Diffusion, are designed to generate carbon copies of the images on which they’ve been trained. In practice, however, they seem to improvise, blending elements within images to create something new — not just nonsensical blobs of color, but coherent images with semantic meaning. This is the “paradox” behind diffusion models, said Giulio Biroli, an AI researcher and physicist at the École Normale Supérieure in Paris: “If they worked perfectly, they should just memorize,” he said. “But they don’t — they’re actually able to produce new samples.”

You may like

Would outsourcing everything to AI cost us our ability to think for ourselves?

Why OpenAI’s solution to AI hallucinations would kill ChatGPT tomorrow

AI can’t solve these puzzles that take humans only seconds

To generate images, diffusion models use a process known as denoising. They convert an image into digital noise (an incoherent collection of pixels), then reassemble it. It’s like repeatedly putting a painting through a shredder until all you have left is a pile of fine dust, then patching the pieces back together. For years, researchers have wondered: If the models are just reassembling, then how does novelty come into the picture? It’s like reassembling your shredded painting into a completely new work of art.

Now two physicists have made a startling claim: It’s the technical imperfections in the denoising process itself that leads to the creativity of diffusion models. In a paper that will be presented at the International Conference on Machine Learning 2025, the duo developed a mathematical model of trained diffusion models to show that their so-called creativity is in fact a deterministic process — a direct, inevitable consequence of their architecture.
By illuminating the black box of diffusion models, the new research could have big implications for future AI research — and perhaps even for our understanding of human creativity. “The real strength of the paper is that it makes very accurate predictions of something very nontrivial,” said Luca Ambrogioni, a computer scientist at Radboud University in the Netherlands.
Mason Kamb, a graduate student studying applied physics at Stanford University and the lead author of the new paper, has long been fascinated by morphogenesis: the processes by which living systems self-assemble.

Sign up for the Live Science daily newsletter now
Get the world’s most fascinating discoveries delivered straight to your inbox.
Contact me with news and offers from other Future brandsReceive email from us on behalf of our trusted partners or sponsorsBy submitting your information you agree to the Terms & Conditions and Privacy Policy and are aged 16 or over.
One way to understand the development of embryos in humans and other animals is through what’s known as a Turing pattern, named after the 20th-century mathematician Alan Turing. Turing patterns explain how groups of cells can organize themselves into distinct organs and limbs. Crucially, this coordination all takes place at a local level. There’s no CEO overseeing the trillions of cells to make sure they all conform to a final body plan. Individual cells, in other words, don’t have some finished blueprint of a body on which to base their work. They’re just taking action and making corrections in response to signals from their neighbors. This bottom-up system usually runs smoothly, but every now and then it goes awry — producing hands with extra fingers, for example.
When the first AI-generated images started cropping up online, many looked like surrealist paintings, depicting humans with extra fingers. These immediately made Kamb think of morphogenesis: “It smelled like a failure you’d expect from a [bottom-up] system,” he said.
AI researchers knew by that point that diffusion models take a couple of technical shortcuts when generating images. The first is known as locality: They only pay attention to a single group, or “patch,” of pixels at a time. The second is that they adhere to a strict rule when generating images: If you shift an input image by just a couple of pixels in any direction, for example, the system will automatically adjust to make the same change in the image it generates. This feature, called translational equivariance, is the model’s way of preserving coherent structure; without it, it’s much more difficult to create realistic images.

You may like

Would outsourcing everything to AI cost us our ability to think for ourselves?

Why OpenAI’s solution to AI hallucinations would kill ChatGPT tomorrow

AI can’t solve these puzzles that take humans only seconds

In part because of these features, diffusion models don’t pay any attention to where a particular patch will fit into the final image. They just focus on generating one patch at a time and then automatically fit them into place using a mathematical model known as a score function, which can be thought of as a digital Turing pattern.
Researchers long regarded locality and equivariance as mere limitations of the denoising process, technical quirks that prevented diffusion models from creating perfect replicas of images. They didn’t associate them with creativity, which was seen as a higher-order phenomenon.
They were in for another surprise.
Made locally
Kamb started his graduate work in 2022 in the lab of Surya Ganguli, a physicist at Stanford who also has appointments in neurobiology and electrical engineering. OpenAI released ChatGPT the same year, causing a surge of interest in the field now known as generative AI. As tech developers worked on building ever-more-powerful models, many academics remained fixated on understanding the inner workings of these systems.

Mason Kamb (left) and Surya Ganguli found that the creativity in diffusion models is a consequence of their architecture. (Image credit: Charles Yang)
To that end, Kamb eventually developed a hypothesis that locality and equivariance lead to creativity. That raised a tantalizing experimental possibility: If he could devise a system to do nothing but optimize for locality and equivariance, it should then behave like a diffusion model. This experiment was at the heart of his new paper, which he wrote with Ganguli as his co-author.
Kamb and Ganguli call their system the equivariant local score (ELS) machine. It is not a trained diffusion model, but rather a set of equations which can analytically predict the composition of denoised images based solely on the mechanics of locality and equivariance. They then took a series of images that had been converted to digital noise and ran them through both the ELS machine and a number of powerful diffusion models, including ResNets and UNets.
The results were “shocking,” Ganguli said: Across the board, the ELS machine was able to identically match the outputs of the trained diffusion models with an average accuracy of 90% — a result that’s “unheard of in machine learning,” Ganguli said.
The results appear to support Kamb’s hypothesis. “As soon as you impose locality, [creativity] was automatic; it fell out of the dynamics completely naturally,” he said. The very mechanisms which constrained diffusion models’ window of attention during the denoising process — forcing them to focus on individual patches, regardless of where they’d ultimately fit into the final product — are the very same that enable their creativity, he found. The extra-fingers phenomenon seen in diffusion models was similarly a direct by-product of the model’s hyperfixation on generating local patches of pixels without any kind of broader context.
Experts interviewed for this story generally agreed that although Kamb and Ganguli’s paper illuminates the mechanisms behind creativity in diffusion models, much remains mysterious. For example, large language models and other AI systems also appear to display creativity, but they don’t harness locality and equivariance.
“I think this is a very important part of the story,” Biroli said, “[but] it’s not the whole story.”
Creating creativity
For the first time, researchers have shown how the creativity of diffusion models can be thought of as a by-product of the denoising process itself, one that can be formalized mathematically and predicted with an unprecedentedly high degree of accuracy. It’s almost as if neuroscientists had put a group of human artists into an MRI machine and found a common neural mechanism behind their creativity that could be written down as a set of equations.

RELATED STORIES

—Why do AI chatbots use so much energy?
—AI slop is on the rise — what does it mean for how we use the internet?
—Scientists just developed a new AI modeled on the human brain — it’s outperforming LLMs like ChatGPT at reasoning tasks
The comparison to neuroscience may go beyond mere metaphor: Kamb and Ganguli’s work could also provide insight into the black box of the human mind. “Human and AI creativity may not be so different,” said Benjamin Hoover, a machine learning researcher at the Georgia Institute of Technology and IBM Research who studies diffusion models. “We assemble things based on what we experience, what we’ve dreamed, what we’ve seen, heard or desire. AI is also just assembling the building blocks from what it’s seen and what it’s asked to do.” Both human and artificial creativity, according to this view, could be fundamentally rooted in an incomplete understanding of the world: We’re all doing our best to fill in the gaps in our knowledge, and every now and then we generate something that’s both new and valuable. Perhaps this is what we call creativity.
Original story reprinted with permission from Quanta Magazine, an editorially independent publication supported by the Simons Foundation.

Webb Wright

Webb Wright is a journalist based in Brooklyn, New York, who writes about technology and the mind. He’s an alumnus of the Columbia University Graduate School of Journalism and a former Ferriss – UC Berkeley Psychedelic Journalism fellow.

You must confirm your public display name before commenting

Please logout and then login again, you will then be prompted to enter your display name.

Would outsourcing everything to AI cost us our ability to think for ourselves?

Why OpenAI’s solution to AI hallucinations would kill ChatGPT tomorrow

AI can’t solve these puzzles that take humans only seconds

Scientists asked ChatGPT to solve a math problem from more than 2,000 years ago — how it answered it surprised them

AI could use online images as a backdoor into your computer, alarming new study suggests

AI outsmarted 30 of the world’s top mathematicians at secret meeting in California

Latest in Artificial Intelligence

Why OpenAI’s solution to AI hallucinations would kill ChatGPT tomorrow

Scientists asked ChatGPT to solve a math problem from more than 2,000 years ago — how it answered it surprised them

‘There’s no shoving that genie back in the bottle’: Readers believe it’s too late to stop the progression of AI

AI could use online images as a backdoor into your computer, alarming new study suggests

Why do AI chatbots use so much energy?

AI slop is on the rise — what does it mean for how we use the internet?

Latest in News

Is acetaminophen safe in pregnancy? Here’s what the science says.

Mysterious cosmic explosion can’t be explained, scientists say

Science news this week: A breakthrough cure for Huntington’s disease and a fast-growing black hole that breaks physics

Amazon rainforest trees are resisting climate change by getting fatter from CO2 in the atmosphere

Scientists unveil world’s first quantum computer built with regular silicon chips

‘If there is a space race, China’s already winning it’: NASA unlikely to bring Mars samples back to Earth before China does, experts say

LATEST ARTICLES

Science history: Rosetta stone is deciphered, opening a window into ancient Egyptian civilization — Sept. 27, 1822

Scientists asked ChatGPT to solve a math problem from more than 2,000 years ago — how it answered it surprised them

Science news this week: A breakthrough cure for Huntington’s disease and a fast-growing black hole that breaks physics

It’s official: Humans have found 6,000 planets beyond our solar system

Why does medicine taste bad?

Live Science is part of Future US Inc, an international media group and leading digital publisher. Visit our corporate site.

Contact Future’s experts

Terms and conditions

Privacy policy

Cookies policy

Accessibility Statement

Advertise with us

Web notifications

Editorial standards

How to pitch a story to us

Future US, Inc. Full 7th Floor, 130 West 42nd Street,

Please login or signup to comment

Please wait…