79
How to block AI Crawler Bots using robots.txt file
(www.cyberciti.biz)
Privacy has become a very important issue in modern society, with companies and governments constantly abusing their power, more and more people are waking up to the importance of digital privacy.
In this community everyone is welcome to post links and discuss topics related to privacy.
[Matrix/Element]Dead
much thanks to @gary_host_laptop for the logo design :)
Did you ban it in your humans.txt too?
humans typically don't visit [website]/fdfjsidfjsidojfi43j435345 when there's no button that links to it
I used to do this on one of my sites that was moderately popular in the 00's. I had a link hidden via javascript, so a user couldn't click it (unless they disabled javascript and clicked it), though it was hidden pretty well for that too.
IP hits would be put into a log and my script would add a /24 of that subnet into my firewall. I allowed specific IP ranges for some search engines.
Anyway, it caught a lot of bots. I really just wanted to stop automated attacks and spambots on the web front.
I also had a honeypot port that basically did the same thing. If you sent packets to it, your /24 was added to the firewall for a week or so. I think I just used netcat to add to yet another log and wrote a script to add those /24's to iptables.
I did it because I had so much bad noise on my logs and spambots, it was pretty crazy.
This thread has provided genius ideas I somehow never thought of, and I'm totally stealing them for my sites lol.
I LOVE VISITING FDFJSIDFJSIDOJFI435345 ON HUMAN WEBSITES, IT IS ONE OF MY FAVORITE HUMAN HOBBIES. ~~🤖~~👨