logoalt Hacker News

freedombenlast Thursday at 6:49 PM7 repliesview on HN

The cybersecurity angle is interesting, because in my experience OpenAI stuff has gotten terrible at cybersecurity because it simply refuses to do anything that can be remotely offensive (as in the opposite of "defensive"). I really thought we as an industry had learned our lesson that blocking "good guys" (aka white-hats) from offensive tools/capabilities only empowers the gray-hat/black-hats and puts us at a disadvantage. A good defense requires some offense. I sure hope they change that.


Replies

tptaceklast Thursday at 7:38 PM

That's odd, because I'm using plain-old-GPT5 as the backend model for a bunch of offensive stuff and I haven't had any hangups at all. But I'm doing a multi-agent setup where each component has a constrained view of the big picture (ie, a fuzzer agent with tool calls to drive a web fuzzer looking for a particular kind of vulnerability); the high-level orchestration is still mostly human-mediated.

prettyblockslast Thursday at 8:44 PM

ChatGPT is very happy to help me with offensive tasks. Codex is as well.

show 1 reply
mapontoseventhslast Thursday at 6:51 PM

The article mentions that more permissive models would be invite only. I think it's a solid approach, as long as they don't make getting one of those invites too difficult.

> "In parallel, we’re piloting invite-only trusted access to upcoming capabilities and more permissive models for vetted professionals and organizations focused on defensive cybersecurity work. We believe that this approach to deployment will balance accessibility with safety."

show 1 reply
JacobAsmuthlast Thursday at 6:57 PM

So in general you think that making frontier AI models more offensive in black hat capabilities will be good for cybersecurity?

show 5 replies
hhhlast Thursday at 6:56 PM

I use openai models every day for offensive work. haven’t had a problem in a long time

nikanjlast Thursday at 7:06 PM

OpenAI is really weird about this stuff. I tried to get good minor chord progression out of chatgpt, but it kept running into guardrails and giving Very Serious Warnings. It felt as if there’s just a dumb keyword filter in there, and getting any amounts of verboted words will kill the entire prompt

show 1 reply
julienfr112last Thursday at 7:49 PM

More generaly, GPT is being heavily neuterd : For exemple I tried to make it rebuild codex itself. It start to answer, then delete the code and go "I'm not to answer that". As if building codex inside codex is a way to terminator and co..