logoalt Hacker News

zozbot234yesterday at 1:59 AM1 replyview on HN

I don't think anyone knows for sure how much mileage/scalability LLMs have. Given what we do know, I suspect if you can afford to spend more compute on even longer training runs, you can still get much better results compared to SOTA, even for "simple" domains like text/language.


Replies

airstrikeyesterday at 3:10 AM

I think we're pretty much out of "spend more compute on even longer training runs" atp.