logoalt Hacker News

FergusArgyllyesterday at 8:37 PM8 repliesview on HN

A new pretrain would definitely get more than a .1 version bump & would get a whole lot more hype I'd think. They're expensive to do!


Replies

caconym_yesterday at 9:10 PM

Releasing anything as "GPT-6" which doesn't provide a generational leap in performance would be a PR nightmare for them, especially after the underwhelming release of GPT-5.

I don't think it really matters what's under the hood. People expect model "versions" to be indexed on performance.

ACCount37yesterday at 9:15 PM

Not necessarily. GPT-4.5 was a new pretrain on top of a sizeable raw model scale bump, and only got 0.5 - because the gains from reasoning training in o-series overshadowed GPT-4.5's natural advantage over GPT-4.

OpenAI might have learned not to overhype. They already shipped GPT-5 - which was only an incremental upgrade over o3, and was received poorly, with this being a part of the reason why.

show 1 reply
hannesfuryesterday at 8:57 PM

Maybe they felt the increase in capability is not worth of a bigger version bump. Additionally pre-training isn't as important as it used to be. Most of the advances we see now probably come from the RL stage.

femiagbabiakayesterday at 8:56 PM

Not if they didn't feel that it delivered customer value no? It's about under promising and over delivering, in every instance

jumploopstoday at 1:17 AM

It’s possible they’re using some new architecture to get more up-to-date data, but I think that’d be even more of a headline.

My hunch is that this is the same 5.1 post-training on a new pretrained base.

Likely rushed out the door faster than they initially expected/planned.

OrangeMusictoday at 7:49 AM

Yeah because OpenAI has been great at naming their models so far? ;)

boctoday at 12:05 AM

Maybe the rumors about failed training runs weren't wrong...

redwoodyesterday at 8:57 PM

Not if it underwhelms