LLM's don't really output the same code quality as a human, even on the smallest scale. It's not even close. Maybe you can guide them to refactor their slop up to human-written quality, but then you're still coding. You're just doing it by asking the computer to write something, instead of physically typing the whole thing out with a keyboard.
This just isn’t true at all, with guidance and guard rails they produce much better code than the average developer does. They are only going to get better.
Yeah I also keep thinking this. I don't see LLMs reliably producing code that is up to my standards. Granted I have high standards because I do take pride in producing high quality code (in all manner of metrics). A lot of the time the code works, unfortunately only for the most naive, mechanical definition of "works".