logoalt Hacker News

JKCalhounlast Saturday at 3:41 PM1 replyview on HN

I agree with that. The question I suppose is whether an LLM can detect, perhaps by the question itself, if they are dealing with someone (I hate to say it) "stable".

Anyone asking how to commit suicide, as a recent example, should be an obvious red flag. We can get more nuanced from there.


Replies

tedmistonlast Saturday at 10:13 PM

> The question I suppose is whether an LLM can detect, perhaps by the question itself, if they are dealing with someone (I hate to say it) "stable".

GPT-5 made a major advance on mental health guardrails in sensitive conversations.

https://www.theverge.com/news/718407/openai-chatgpt-mental-h...

https://openai.com/index/strengthening-chatgpt-responses-in-...