Human thoughts are “hidden” in the sense that they exist separately from the text being written. They will correlate somewhat with that text of course, but they aren’t completely determined by it.
The only state for GPT-like models is that which is supplied in the previous text. They don’t have any ‘private’ state at all, not even between one token and the next. This is a very clear difference, and does in both principle and practice constrain their behaviour.
Human thoughts are “hidden” in the sense that they exist separately from the text being written. They will correlate somewhat with that text of course, but they aren’t completely determined by it.
The only state for GPT-like models is that which is supplied in the previous text. They don’t have any ‘private’ state at all, not even between one token and the next. This is a very clear difference, and does in both principle and practice constrain their behaviour.
They can compute a state prior to each generated token and they can choose a token that signal a preservation of this state.