logoalt Hacker News

dvtyesterday at 11:36 PM3 repliesview on HN

> I don't really understand why this type of pattern occurs, where the later words in a sentence don't properly connect to the earlier ones in AI-generated text.

Because AI is not intelligent, it doesn't "know" what it previously output even a token ago. People keep saying this, but it's quite literally fancy autocorrect. LLMs traverse optimized paths along multi-dimensional manifolds and trick our wrinkly grey matter into thinking we're being talked to. Super powerful and very fun to work with, but assuming a ghost in the shell would be illusory.


Replies

Tossrockyesterday at 11:55 PM

> Because AI is not intelligent, it doesn't "know" what it previously output even a token ago.

Of course it knows what it output a token ago, that's the whole point of attention and the whole basis of the quadratic curse.

show 1 reply
Borealidyesterday at 11:41 PM

If all the training data contains semantically-meaningful sentences it should be possible to build a network optimized for generating semantically-meaningful sentence primarily/only.

But we don't appear to have entirely done that yet. It's just curious to me that the linguistic structure is there while the "intelligence", as you call it, is not.

show 4 replies
CamperBob2yesterday at 11:53 PM

Because AI is not intelligent, it doesn't "know" what it previously output even a token ago.

You have no idea what you're talking about. I mean, literally no idea, if you truly believe that.

show 1 reply