I have read comments about this on X, here, and other places, yet I have ever seen there be proof this is an actual productivity boost.
I use Claude Opus (4.5, 4.6) all the time and catch it making making subtle mistakes, all the time.
Are you really being more productive (let’s say 3x times more), or just feel that way because you are constantly prompting Claude?
Maybe I’m wrong, but I don’t buy it.
I agree. The code despite detailed spec reveals bugs and edge cases upon inspection.
I'm talking Claude Opus 4.6 here.
typical experience when only using one foundational model TBH. results are much better if you let different models review each other.
the bottleneck now is testing. that isn't going away anytime soon, it'll get much worse for a bit while models are good at churning code out that's slightly wrong or technically correct, but solving a different problem than intended; it's going to be a relatively short lived situation I'm afraid until the industry switches to most code being written for serving agents instead of humans.
i really don't understand why people keep thinking this. i'm easily 10x more productive since Claude Code came out. it's insane how much stuff you can build quickly, especially on personal projects.
> I use Claude Opus (4.5, 4.6) all the time and catch it making making subtle mistakes, all the time.
Didn't we make subtle mistakes without AI?
Why did we spend so much time debugging and doing code reviews?
> Are you really being more productive (let’s say 3x times more)
At least 2x more productive, and that's huge.