logoalt Hacker News

stkaiyesterday at 6:56 PM4 repliesview on HN

Would love to find out they're overfitting for pelican drawings.


Replies

andy_pppyesterday at 7:16 PM

Yes, Racoon on a unicycle? Magpie on a pedalo?

show 3 replies
theanonymousoneyesterday at 9:44 PM

Even if not intentionally, it is probably leaking into training sets.

fragmedeyesterday at 7:54 PM

The estimation I did 4 months ago:

> there are approximately 200k common nouns in English, and then we square that, we get 40 billion combinations. At one second per, that's ~1200 years, but then if we parallelize it on a supercomputer that can do 100,000 per second that would only take 3 days. Given that ChatGPT was trained on all of the Internet and every book written, I'm not sure that still seems infeasible.

https://news.ycombinator.com/item?id=45455786

show 2 replies