So ... is there any reason I shouldn't put a nest of meaningless pages on my web site in robots.txt as forbidden, link them into web pages in a way that should be invisible to (or obviously ignored by) humans and un-followed by browsers, and then put URLs for some of the deeper pages into fail2ban? It feels like this should allow human traffic, allow valid search crawlers, and blackhole AI crawlers.
=> More informations about this toot | More toots from elb@social.sdf.org
@elb clever idea to use fail2ban as an enforcement mechanism for robots.txt!
=> More informations about this toot | More toots from aburka@hachyderm.io
@aburka
Ooh, I've used fail2ban for ssh for something like a decade, didn't occur to me that it could score off of other log files, neat!
@elb
=> More informations about this toot | More toots from eichin@mastodon.mit.edu This content has been proxied by September (3851b).Proxy Information
text/gemini