Back to Blog
·3 min read
Edit

Temperature

AI Engineering

You ask an LLM a question and the response arrives perfectly formed precise, broad, and entirely unsurprising. It reads like it was written by committee and approved by legal.

It's like walking past an astroturf garden. The grass is immaculate. Uniform. Technically flawless. And yet it triggers nothing no warmth of a freshly cut meadow, no pride of a well-kept lawn. Something is a little amiss. It's too perfect to feel real. That's the default LLM experience. Predictable by design.

This predictability is good as the alternative doesn't even bear thinking about in relation to the volume of AI slop out there. LLMs generate text by predicting the next token according to a probability distribution. Basically, each candidate to be the next word is assigned a probability. The model then picks from the top of that distribution. The most probable token wins, almost every time.

The result? Safe, high-confidence outputs. The kind of text that's never wrong but rarely surprising.

So does that mean LLMs can't contribute to genuinely novel thinking? Not quite. When paired with a skilled human, they're powerful collaborative partners, amplifying ideas, stress-testing angles, filling gaps. But in isolation, a model trained on existing data and biased toward the most likely next word is not naturally wired for novelty.

Which brings us to the temperature parameter of an LLM.

People often call temperature the "creativity dial." That's a tempting label, but it's not quite right. Temperature doesn't make the model necessarily think differently. It changes how it chooses. At low temperature, the model doubles down on its highest-confidence predictions. The output is tight, deterministic, and safe. At high temperature, the probability distribution flattens less likely tokens get a real shot at being selected. The output becomes more variable, more unexpected, occasionally more interesting.

Nutshell: temperature doesn't increase creativity. It increases randomness.

And randomness, used deliberately, can be a brute force creative tool. This is why I think of temperature not as a creativity setting, but as a wildcard dial.

Creativity is a human skill connecting dots, reading context, making judgement calls about what's good. An LLM can't do that. But what it can do, when you turn up the temperature, is throw you tokens you wouldn't have expected. If you're stuck in a loop of predictable outputs and want to shake something loose, the wildcard dial is there for a reason. When to use it:

Low temperature (0–0.3): When you need precision, code generation, factual summaries, structured data extraction.

Medium temperature (0.4–0.7): General-purpose writing and conversation. The default sweet spot.

High temperature (0.8–1.2+): When you're brainstorming, exploring, or deliberately looking for the unexpected. When you want the wildcard.

Next time your LLM outputs feel like astroturf; neat, predictable, emotionally flat turn up the wildcard dial or needs some venom injected. You might not use what comes back. But it might be exactly the spark that leads you somewhere new.

Erik Cavan

Erik Cavan

Applied AI

Share: