logoalt Hacker News

antdkeyesterday at 9:38 PM4 repliesview on HN

Well, imagine this was controlling a weapon.

“Should I eliminate the target?”

“no”

“Got it! Taking aim and firing now.”


Replies

bigstrat2003yesterday at 9:40 PM

It is completely irresponsible to give an LLM direct access to a system. That was true before and remains true now. And unfortunately, that didn't stop people before and it still won't.

show 1 reply
nielsoleyesterday at 9:39 PM

Shall I open the pod bay doors?

nvchyesterday at 9:54 PM

"Thinking: the user recognizes that it's impossible to guarantee elimination. Therefore, I can fulfill all initial requirements and proceed with striking it."

verdvermyesterday at 9:40 PM

That's why we keep humans in the loop. I've seen stuff like this all the time. It's not unusual thinking text, hence the lack of interestingness

show 1 reply