In all these discussions there seems to be an inverse correlation between how well people understand what an LLM does and how much they believe it thinks.
If you don't understand what an LLM does – that it is a machine generating a statistically probable token given a set of other tokens – you have a black box that often sounds smart, and it's pretty natural to equate that to thinking.
"Next token prediction" is not an answer. It's mental shortcut. An excuse not to think about the implications. An excuse a lot of people are eager to take.
First, autoregressive next token prediction can be Turing complete. This alone should give you a big old pause before you say "can't do X".
Second, "next token prediction" is what happens at an exposed top of an entire iceberg worth of incredibly poorly understood computation. An LLM is made not by humans, but by an inhuman optimization process. No one truly "understands" how an LLM actually works, but many delude themselves into thinking that they do.
And third, the task a base model LLM is trained for - what the optimization process was optimizing for? Text completion. Now, what is text? A product of human thinking expressed in natural language. And the LLM is forced to conform to the shape.
How close does it get in practice to the original?
Not close enough to a full copy, clearly. But close enough that even the flaws of human thinking are often reproduced faithfully.