logoalt Hacker News

irthomasthomasyesterday at 9:32 PM4 repliesview on HN

Isn't this proof that LLMs still don't really generalize beyond their training data?


Replies

adastra22yesterday at 11:23 PM

LLMs are very good at generalizing beyond their training (or context) data. Normally when they do this we call it hallucination.

Only now we do A LOT of reinforcement learning afterwards to severely punish this behavior for subjective eternities. Then act surprised when the resulting models are hesitant to venture outside their training data.

show 1 reply
Zambyteyesterday at 10:49 PM

I wonder how they would behave given a system prompt that asserts "dogs may have more or less than four legs".

show 1 reply
CamperBob2yesterday at 10:17 PM

They do, but we call it "hallucination" when that happens.

Rover222yesterday at 9:43 PM

Kind of feels that way