logoalt Hacker News

notachatbot123today at 9:25 AM1 replyview on HN

The article is about AI web crawlers. How can your tool help and how would one set it up for this specific context?


Replies

reconnectingtoday at 9:32 AM

I don't see how an AI crawler is different from any others.

The simplest approach is to count the UA as risky or flag multiple 404 errors or HEAD requests, and block on that. Those are rules we already have out of the box.

It's open source, there's no pain in writing specific rules for rate limiting, thus my question.

Plus, we have developed a dashboard for manually choosing UA blocks based on name, but we're still not sure if this is something that would be really helpful for website operators.

show 1 reply