logoalt Hacker News

latexrtoday at 10:28 AM1 replyview on HN

An LLM can always output steps, but it doesn’t mean they are true, they are great at making up bullshit.

When the “how many ‘r’ in ‘strawberry’” question was all the rage, you could definitely get LLMs to explain the steps of counting, too. It was still wrong.


Replies

simianwordstoday at 10:30 AM

can you provide a single example now with gpt 5.4 thinking that makes up things in steps? lets try to reproduce it.

show 1 reply