← All news

AI Will Get Creative the Same Way Nature Did: Selection Pressure

There is a familiar complaint about large language models: they are not creative. The mistake is assuming the model is failing. It is doing what it was built to do.

There is a familiar complaint people keep repeating about large language models: they are not creative. They can remix, they can imitate, they can generate, but they do not truly invent. And if you spend enough time around these systems, you can see why the complaint feels right. Ask an LLM for a brilliant new idea and it will often give you something that looks polished, even clever, but somehow expected. Like it has been pre-approved by a committee. The mistake, though, is assuming the model is failing at its job. It is not. It is doing exactly what it was built to do: predict what comes next, based on what it has seen before. That is not a moral limitation. It is an architectural one.

When people talk about creativity, they often imagine a mysterious spark. The moment Mozart hears a melody in his head. The moment a designer suddenly sees the logo that "just works." The moment a founder connects two ideas that did not belong together yesterday but somehow make perfect sense today. But if you look closely, creativity is rarely a lightning strike. It is a process. It is variation and selection. It is throwing things at the wall, noticing what sticks, and then doing it again, slightly better, slightly sharper, slightly more you. What looks like magic from the outside is often repetition with taste.

This is where the temperature slider enters the conversation, because it is the one thing people can point to and say, "See? That is creativity." Turn it up and the model gets weird. Turn it down and it gets safe. But temperature is not creativity, it is controlled randomness. It is the model being less certain about what the next word should be, so it samples from a wider set of possibilities. Sometimes that gives you something fresh. Often it gives you noise. The model does not know which is which. Humans do not just generate. Humans judge. They have an internal scoring system shaped by culture, status, pain, reward, ambition, and embarrassment. A model, left alone, has none of that. It can produce infinite options and still have no idea which one is worth keeping.

But there is a deeper layer to this debate that people almost never mention, and it is the one that matters most: creative for whom? We talk about creativity as if it is a universal property, like gravity, something that exists inside the output itself. But creativity is relational. It lives in the reaction. A logo is not creative in isolation. A slogan is not creative in isolation. Even a scientific theory is not "creative" floating alone in space. Something becomes creative when it surprises the right observer in the right context, while still feeling correct. That means the same output can be boring to one group and brilliant to another. And it also means the same AI system can look "uncreative" when judged by the wrong audience.

This is where the idea of selection pressure becomes the real story. Nature did not design the eye by having a single organism invent it. Nature produced endless variation, most of it useless, some of it slightly better, and then the environment did the judging. Survival became the filter. The result looks like genius, but the mechanism is brutally simple. You do not need a divine spark. You need a loop that rewards what works.

Now imagine we stop trying to force LLMs to be "creative" in the romantic way people talk about artists, and instead we build the loop. You generate ten logo variations, or fifty, or five hundred. Then you do not sit in a room arguing about which one is best. You run a micro-panel. A, B, C, or D. Quick choices. Real humans. Not a design committee, not your co-founder's opinion, not the loudest person in the Slack channel. Just preference signals, at scale. And then you add context: who are these people, what do they like, what demographic are they in, what brands do they already trust, what aesthetic world do they live in. Suddenly the model is not just generating. It is generating into an environment that can judge.

This works beautifully when you are building software for humans. But the moment you build software for software, the entire scoring system changes. If the customer is not a person but another agent, then "creative" does not mean delightful. It means effective. It means cheaper, faster, more robust, more correct. It means something like: does this solution reduce compute, reduce latency, increase success rate, reduce risk, improve retrieval, avoid failure modes, generalize better. That kind of creativity does not need a focus group. It needs benchmarks. It needs adversarial tests. It needs hard feedback loops that are almost clinical. The irony is that agent-to-agent creativity might scale faster than human-facing creativity, because software can judge software at insane speed. It does not get tired. It does not get moody. It does not change its mind because the last option "felt nicer."

So you end up with two different worlds of AI creativity. Human-facing creativity is taste-driven and emotional, and it needs humans in the loop. Agent-facing creativity is performance-driven and mechanical, and it needs metrics in the loop. Both are selection pressure. They just use different environments as the judge.

And then there is the third world, the one that looks like creativity but is closer to science. Science and creativity are not the same thing, but they rhyme. The loop is still there. The only difference is the feedback signal. In design, the feedback is preference. In science, the feedback is truth. You do not ask a panel which hypothesis is nicer. You run an experiment. You try to break the idea. You measure reality and see if it agrees. It is still generate, test, select, mutate, repeat. Just with higher stakes and harsher scoring.

This is why the next leap in "AI creativity" will not come from better prompts or bigger context windows. It will come from letting software run more experiments. The more trials it can do, the more variation it can explore, and the faster it can update its internal beliefs about what works. If you can run ten thousand simulated chemistry experiments overnight, you are not doing creativity in the human sense. You are doing invention in the most literal sense. The system is walking through possibility space and letting reality act as the filter.

And we are not far away from this. We are already watching it happen in pharma, where the loop is tightening fast: generate candidate molecules, predict properties, test them, update the model, repeat. The same pattern is emerging in materials science, battery chemistry, protein engineering, and even pure math, where "creativity" looks like finding new paths through proofs and new ways to compress complexity. As soon as the experiments become cheaper, faster, and more automated, the rate of invention jumps. Not because the model becomes more soulful, but because the environment becomes more responsive.

This is the point people miss when they argue about whether LLMs are creative. They are staring at the generator and ignoring the filter. They are judging the output and ignoring the loop. Creativity is not a personality trait. It is an evolutionary process. And once we build the right selection pressure, whether it is human taste, agent metrics, or physical experiments, AI will not need to "feel" creative. It will simply become inventive, because it will be forced to earn its survival in the only way that matters: by producing things that work.

First published on Medium.