This isn't really true though. Pre-training for coding models is just a mass of scraped source-code, but post-training is more than simply generating compiling code. It includes extensive reinforcement learning of curated software-engineering tasks that are designed to teach what high quality code looks like, and to improve abilities like debugging, refactoring, tool use, etc.
There’s no objective measurement for high quality code, so I don’t think model creators are going to be particularly good at screening for it.
Well and also a lot of Claude Code users data as well. That telemetry is invaluable.