logoalt Hacker News

poormanlast Sunday at 8:52 PM2 repliesview on HN

As we saw with GPT-5 the RL technique of training doesn't scale forever


Replies

energy123last Monday at 12:12 PM

Unless GPT-5 is 30% cheaper to run than o3. Then it's scaling brilliantly given the small gap between release dates. People are really drawing too many conclusions from too little information.

oezilast Sunday at 9:19 PM

I meant scaling the base training before RL.