Here's one. Given a conversation history made of n sequential tokens S1, S2, ..., Sn, an LLM will generate the next token using an insanely complicated model we'll just call F:
S(n+1) = F(S1, S2, ..., Sn)
As for me, I'll often think of my next point, figure out how to say that concept, and then figure out the right words to connect it where the conversation's at right then. So there's one function, G, for me to think of the next conversational point. And then another, H, to lead into it. S(n+100) = G(S1, S2, ..., Sn)
S(n+1) = G(S1, S2, ..., Sn, S(n+100))
And this is putting aside how people don't actually think in tokens. And some people don't always have an internal monologue (I rarely do when doing math).