This is an article from 2024, when open weights models like llama were only beginning to emerge. With those you basically cannot reliably do any detection (as the authors admit by the end).
Which is really boiling down to text having statistically very similar properties to human generated one. Introduce a more motivated attacker and the text would be indistinguishable from real (with occasional typos, no use of "delve", "it's not x its y", emdashes and so on).
It really is a lost battle: you cannot embed extra information in the text that will survive even basic postprocessing (in contrast to, say, steganography)
It sounds like a "cursed problem". Are there any contemporary techniques that show any promise?
Ultimately it shouldn’t be too surprising that the machine that works by generating the most statistically likely text, generates text that’s statistically identical to human-generated text