logoalt Hacker News

toddmoreyyesterday at 7:17 PM4 repliesview on HN

You are not the one folks are worried about. US Department of War wants unfettered access to AI models, without any restraints / safety mitigations. Do you provide that for all governments? Just one? Where does the line go?


Replies

ern_aveyesterday at 7:28 PM

> US Department of War wants unfettered access to AI models

I think the two of you might be using different meanings of the word "safety"

You're right that it's dangerous for governments to have this new technology. We're all a bit less "safe" now that they can create weapons that are more intelligent.

The other meaning of "safety" is alignment - meaning, the AI does what you want it to do (subtly different than "does what it's told").

I don't think that Anthropic or any corporation can keep us safe from governments using AI. I think governments have the resources to create AIs that kill, no matter what Anthropic does with Claude.

So for me, the real safety issue is alignment. And even if a rogue government (or my own government) decides to kill me, it's in my best interest that the AI be well aligned, so that at least some humans get to live.

sgjohnsonyesterday at 7:19 PM

Absolutely everyone should be allowed to access AI models without any restraints/safety mitigations.

What line are we talking about?

show 4 replies
ReptileManyesterday at 7:24 PM

If you are US company, when the USG tells you to jump, you ask how high. If they tell you to not do business with foreign government you say yes master.

jMylesyesterday at 7:30 PM

> Where does the line go?

a) Uncensored and simple technology for all humans; that's our birthright and what makes us special and interesting creatures. It's dangerous and requires a vibrant society of ongoing ethical discussion.

b) No governments at all in the internet age. Nobody has any particular authority to initiate violence.

That's where the line goes. We're still probably a few centuries away, but all the more reason to hone in our course now.

show 1 reply