When we talk to ChatGPT, it feels like we’re chatting with something that thinks. But in reality, there’s no actual thought going on. What happens is that your input is broken into small units called tokens, and the model predicts the next most likely token based on patterns it has learned from massive amounts of text. It doesn’t understand meaning the way humans do, it’s simply mapping probabilities.

At the starting line of TNF100 2012, we don’t know every step ahead, just like how neural networks don’t really ‘think,’ they just follow patterns.
The magic lies in the transformer architecture, especially attention. Attention allows the model to weigh which parts of your input are more important in predicting the next word, kind of like how a runner decides which signals from their body to prioritize during a long run: pace, breath, or fatigue. Each prediction step is just math, but when you string billions of these steps together, the result feels like coherent conversation.

Just as kids learn by seeing and repeating, neural networks ‘learn’ patterns, but they don’t truly think.
This is why it seems like ChatGPT is thinking.
IT ISN’T.
What you’re seeing is the illusion of thought, language modeled so well that it mirrors reasoning. Much like a pacer in a marathon doesn’t run the race for you but keeps you in rhythm, ChatGPT doesn’t think but generates responses that keep the dialogue flowing. It’s not a mind; it’s a tool powered by data, math, and prediction.
Instead of trying to produce a program to simulate the adult mind, why not rather try to produce one which simulates the child’s?
Alan Turing