> I don't really understand why this type of pattern occurs, where the later words in a sentence don't properly connect to the earlier ones in AI-generated text.
Because AI is not intelligent, it doesn't "know" what it previously output even a token ago. People keep saying this, but it's quite literally fancy autocorrect. LLMs traverse optimized paths along multi-dimensional manifolds and trick our wrinkly grey matter into thinking we're being talked to. Super powerful and very fun to work with, but assuming a ghost in the shell would be illusory.
If all the training data contains semantically-meaningful sentences it should be possible to build a network optimized for generating semantically-meaningful sentence primarily/only.
But we don't appear to have entirely done that yet. It's just curious to me that the linguistic structure is there while the "intelligence", as you call it, is not.
Because AI is not intelligent, it doesn't "know" what it previously output even a token ago.
You have no idea what you're talking about. I mean, literally no idea, if you truly believe that.
> Because AI is not intelligent, it doesn't "know" what it previously output even a token ago.
Of course it knows what it output a token ago, that's the whole point of attention and the whole basis of the quadratic curse.