logoalt Hacker News

ironboundyesterday at 1:17 PM5 repliesview on HN

LLM's are why we're in this mess, they can't do math or count r's


Replies

gordonhartyesterday at 1:29 PM

Modern reasoning models are actually pretty good at arithmetic and almost certainly would have caught this error if asked.

Source: we benchmark this sort of stuff at my company and for the past year or so frontier models with a modest reasoning budget typically succeed at arithmetic problems (except for multiplication/division problems with many decimal places, which this isn't).

show 1 reply
literalAardvarkyesterday at 2:01 PM

They can't do math?

ChatGPT 5.2 has recently been churning through unsolved Erdös problems.

I think right now one is partially validated by a pro and the other one I know of is "ai-solved" but not verified. As in: we're the ones who can't quite keep up.

https://arxiv.org/abs/2601.07421

And the only reason they can't count Rs is that we don't show them Rs due to a performance optimization.

show 1 reply
nine_kyesterday at 1:36 PM

An LLM usually has a powerful digital computer right in its disposal, and could use it as a tool to do precise calculations.

brookstyesterday at 1:46 PM

More accurate to say they can’t see r’s. They process language but not letters.

UqWBcuFx6NV4ryesterday at 1:57 PM

Yes, yes. We’ve all seen the same screenshots. Very funny.

Those of us who don’t base our technical understandings on memes are well aware of the tooling at the disposal of all modern reasoning models gives them the capability to do such things.

Please don’t bring the culture war here.