Did you know that there’s a loaded dice inside ChatGPT? π²
β¨ Part 2/5 of the Mini-Series: The Creative Shell β What Makes LLMs “Creative”?
see previous part here
Every time an LLM generates text, it’s rolling a dice π².
But here’s the thing, the dice is loaded π± !
In Part 0 & 1, we explored how temperature scales logits to control creativity. Today we’re going one level deeper: how those scaled logits become the probabilities the model actually samples from.
When the model computes logits for “The sky is ⦔, you might get:
- “the” β
β93.7 - “blue” β
β94.3 - “falling” β
β94.5 - “a” β
β94.8
These are just raw scores. You can’t sample from them yet.
Here “Softmax” comes to play (see image π). It does two elegant things:

1οΈβ£ Takes e^(each scaled logit) β converts negatives to positives
2οΈβ£ Divides by the sum β forces everything to add up to 100%
Now we have a probability distributionβthe model’s loaded dice! π²
Looking at the image below, notice the effect of the Temperature:

πΉ LOW TEMPERATURE (0.5) β‘οΈ Peaked Distribution:
“the”: 57.30% β Clear winner
“blue”: 18.20%
“falling”: 13.09%
“a”: 6.36%
Result? More than 50% chance of picking “the.”. Leads to predictable model.
πΉ HIGH TEMPERATURE (2.0) β‘οΈ Flat Distribution:
- “the”:
1.21% - “blue”:
0.90% - “falling”:
0.83% - “a”:
0.70%
Result? No dominant choice. Many words compete. Leads to unpredictable model (sometimes chaotic).
π§ The mind-bending part is that the model never “chooses” words deterministically. It:
1οΈβ£ Creates a dice π² (probability distribution) using the softmax function
2οΈβ£ Randomly samples from it
Think of it like this:
– Low temp = heavily weighted dice (predictable rolls)
– High temp = balanced dice (wild rolls)
That’s why:
β
Same prompt β same output
β
There’s randomness in every response
β
Temperature doesn’t make the model “smarter”βit changes how the dice is weighted
The dice is always there. Temperature just loads it differently.
In part 3, we’ll shape and carve that dice even further
π€ What’s been your most surprising “dice roll” from an LLM?
β»οΈ Repost if this changed how you think about LLMs