logoalt Hacker News

sjdbbddtoday at 2:21 PM1 replyview on HN

Did the author do any audit on correctness? Anytime I let the LLM rip it makes mistakes. Most of the pro AI articles (including agentic coding) like this I read always have this in common:

- Declare victory the moment their initial testing works

- Didn’t do the time intensive work of verifying things work

- Author will personally benefit from AI living up to the hype they’re writing about

In a lot of the authors examples (especially with booking), a single failure would be extremely painful. I’d still want to pay knowing this is not likely to happen, and if it does, I’ll be compensated accordingly.


Replies

afro88today at 5:34 PM

Would love to know this too. When he talks about letting clawdbot catch promises and appointments in his texts, how many of those get missed? How many get created incorrectly? Absolutely not none. But maybe the numbers work compared to how bad he was at it manually?