> Of course it knows what it output a token ago...
It doesn't know anything. It has a bunch of weights that were updated by the previous stuff in the token stream. At least our brains, whatever they do, certainly don't function like that.
i dont think this is a meaningful distinction.
it knows the past tokens because theyre part of the input for predicting the next token. its part of the model architecture that it knows it.
if that isnt knowing, people dont know how to walk, only how to move limbs, and not even that, just a bunch of neurons firing
Wait till you learn how human memory works.
Every time you recall a memory it is modified, every time you verbalise a memory it is modified even more so.
Eye-witness accounts are notoriously unreliable, people who witness the same events can have shockingly differing versions.
Memories are modified when new information, real or fabricated, is added.
It’s entirely possible to convince people to recall events that never occurred.
Which of your memories are you certain are of real occurrences, or memories of dreams?
I don't know anything (or even much) about how our brains function, but the idea of a neuron sending an electrical output when the sum of the strengths of its inputs exceeds some value seems to be me like "a bunch of weights" getting repeatedly updated by stimulus.
To you it might be obvious our brains are different from a network of weights being reconfigured as new information comes in; to me it's not so clear how they differ. And I do not feel I know the meaning of the word "know" clearly enough to establish whether something that can emit fluent text about a topic is somehow excluded from "knowing" about it through its means of construction.