logoalt Hacker News

gruezyesterday at 9:19 PM2 repliesview on HN

>You should check out "model collapse". It seems that an abundance of content, that is more and more AI generated these days, may not be a viable option.

Doom-saying about "model collapse" is kind of funny when OpenAI and Anthropic are mad at Chinese model makers for "distilling" their models, ie. using their outputs to train their own models.


Replies

HWR_14today at 12:34 AM

Totally different use cases. If you have nothing, getting 90% of a SOTA model is very valuable. If you have a SOTA model, it's just a worse model.

quikoatoday at 12:03 AM

Isn't there a difference between: distilling specific AI input/output vs scraping whatever random AI output (with unknown input)?