They used to compare to competing models from Anthropic, Google DeepMind, DeepSeek, etc. Seems that now they only compare to their own models. Does this mean that the GPT-series is performing worse than its competitors (given the "code red" at OpenAI)?
The matrix required for a fair comparison is getting too complicated, since you have to compare chat/thinking/pro against an array of Anthropic and Google models.
But they publish all the same numbers, so you can make the full comparison yourself, if you want to.
They are taking a page out of Apple's book.
Apple only compares to themselves. They don't even acknowledge the existence of others.
OpenAI has never compared their models to models from other labs in their blog post. Open literally any past model launch post to see that.
They did compare it to other models: https://x.com/OpenAI/status/1999182104362668275
https://i.imgur.com/e0iB8KC.png