logoalt Hacker News

suddenlybananasyesterday at 6:01 PM1 replyview on HN

They probably had time to toss that example in the training soup.


Replies

AlphaAndOmega0yesterday at 6:07 PM

Previous models from competitors usually got that correct, and the reasoning versions almost always did.

This kind of reflexive criticism isn't helpful, it's closer to a fully generalized counter-argument against LLM progress, whereas it's obvious to anyone that models today can do things they couldn't do six months ago, let alone 2 years back.

show 1 reply