I do wonder what the moat is around this class of products (call it "coding agents").
My intuition is that it's not deep... the differentiating factor is "regular" (non LLM) code which assembles the LLM context and invokes the LLM in a loop.
Claude/Codex have some advantage, because they can RLHF/finetune better than others. But ultimately this is about context assembly and prompting.
There is no moat. It's all prompts. The only potential moat is building your own specialized models using the code your customers send your way I believe.