Extremely cool! I just wish they would also include comparisons to SOTA models from OpenAI, Google, and Anthropic in the press release, so it's easier to know how it fares in the grand scheme of things.
I guess that could be considered comparative advertising then and companies generally try to avoid that scrutiny.
> I just wish they would also include comparisons to SOTA models from OpenAI, Google, and Anthropic in the press release,
Why would they? They know they can't compete against the heavily closed-source models.
They are not even comparing against GPT-OSS.
That is absolutely and shockingly bearish.
The lack of the comparison (which absolutely was done), tells you exactly what you need to know.
They mentioned LMArena, you can get the results for that here: https://lmarena.ai/leaderboard/text
Mistral Large 3 is ranked 28, behind all the other major SOTA models. The delta between Mistral and the leader is only 1418 vs. 1491 though. I *think* that means the difference is relatively small.