logoalt Hacker News

dheerayesterday at 11:49 PM4 repliesview on HN

If I were in charge of an org's cybersecurity I would have AI agents continually trying to attack the systems 24/7 and inform me of successful exploits; it would suck if the major model providers block this type of usage.


Replies

jsheardtoday at 12:21 AM

Judging from the experience of people running bug bounty programs lately, you'd definitely get an endless supply of successful exploit reports. Whether any of them would be real exploits is another question though.

https://daniel.haxx.se/blog/2025/07/14/death-by-a-thousand-s...

netvaruntoday at 12:44 AM

Shameless plug: We're building this. Our goal is to provide AI pentesting agents that run continuously, because the reality is that companies (eg: those doing SOC 2) typically get a point-in-time pentest once a year while furiously shipping code via Cursor/Claude Code and changing infrastructure daily.

I like how Terence Tao framed this [0]: blue teams (builders aka 'vibe-coders') and red teams (attackers) are dual to each other. AI is often better suited for the red team role, critiquing, probing, and surfacing weaknesses, rather than just generating code (In this case, I feel hallucinations are more of a feature than a bug).

We have an early version and are looking for companies to try it out. If you'd like to chat, I'm at [email protected].

[0] https://mathstodon.xyz/@tao/114915606467203078

show 1 reply
cube00today at 12:07 AM

That sounds expensive, those LLM API calls and tokens aren't cheap.

show 2 replies
idontwantthistoday at 1:41 AM

Horizon3 offers this.