logoalt Hacker News

WarmWashtoday at 2:18 PM1 replyview on HN

>What I am saying is that once the high quality training data runs out, it will drop in its capabilities pretty fast.

That's more a misunderstood study that over time turned into a confidently stated fact. Yes, the model collapses if you loop the output to the input. But no, that's not how it's done.

The reality is that all the labs are already using synthetic training data, and have been for at least a year now. It basically turned out to be a non-issue if you have robust monitoring and curation in place for the generated data.


Replies

qseratoday at 2:41 PM

>using synthetic training data

yea, look up how it is done.

It is exactly how a perpetual motion machine scam would project an appearance of working like using a generator to drive a motor, and the motor driving the generator..something that would obscure the fact that there is energy loss happening along the way....

show 1 reply