Hmm, some LLM text is hard to detect, sure.
Some is also horribly easy. If the text is full of:
- Overly positive commentary and encouragement
- Constant use of bullet point lists, bolding and emoji
- This quaint forced 'funniness', like a misplaced attempt at being lighthearted
- A lot of blablah that just missed the point
- Not concise and to the point, but also not super long
Then that really screams ChatGPT to me.
I think it's because this seems to be the default styling of ChatGPT. When people tailor their prompt to be more specific about style it's a lot harder to detect but if they just dump a few lines of instructions about the content into it, this is what you'll get. So the low-effort slop is still pretty easy to detect IMO.
> This quaint forced 'funniness', like a misplaced attempt at being lighthearted
HN always downvotes attempts at humour, be them chatbot or brain generated :)