logoalt Hacker News

vidarhtoday at 2:18 AM0 repliesview on HN

The irony is that while from perfect, an LLM-based fact-checking agent is likely to be far more dilligent (but still needs human review as well) by nature of being trivial to ensure it has no memory of having done a long list of them (if you pass e.g. Claude a long list directly in the same context, it is prone to deciding the task is "tedious" and starting to take shortcuts).

But at the same time, doing that makes it even more likely the human in the loop will get sloppy, because there'll be even fewer cases where their input is actually needed.

I'm wondering if you need to start inserting intentional canaries to validate if humans are actually doing sufficiently torough reviews.