logoalt Hacker News

bambushuyesterday at 9:29 PM1 replyview on HN

This matches my experience exactly. I built a tool that sends code to three different AI models for review because the model that wrote the code can't critique it honestly. It has all the context and actively suppresses objections. The second model, with zero context, immediately finds things the first one rationalized away. Taste isn't just knowing what good looks like, it's being willing to say "this isn't it" to your own work. AI can't do that to itself yet.


Replies

wallstopyesterday at 10:48 PM

I don't think you really need a tool for that, you can just add something like "after the task is finished, have a subagent review the work in an adversarial fashion. If any defects, no matter how small are found, have another subagent implement the findings. Repeat this in a loop until all subagents achieve consensus that the product is of exceptional quality with no defects" or similar to each prompt. Each subagent gets its own, fresh, context window. No tooling required.