I disallow a page in my robots.txt and ip-ban everyone who goes there. Thats pretty effective.
humans typically don’t visit [website]/fdfjsidfjsidojfi43j435345 when there’s no button that links to it
I used to do this on one of my sites that was moderately popular in the 00’s. I had a link hidden via javascript, so a user couldn’t click it (unless they disabled javascript and clicked it), though it was hidden pretty well for that too.
IP hits would be put into a log and my script would add a /24 of that subnet into my firewall. I allowed specific IP ranges for some search engines.
Anyway, it caught a lot of bots. I really just wanted to stop automated attacks and spambots on the web front.
I also had a honeypot port that basically did the same thing. If you sent packets to it, your /24 was added to the firewall for a week or so. I think I just used netcat to add to yet another log and wrote a script to add those /24’s to iptables.
I did it because I had so much bad noise on my logs and spambots, it was pretty crazy.
Imagine posting a rule that says “do not walk on the grass” among other rules and then banning anyone who steps on the grass with the thought process that if they didn’t obey that rule they were likely disobeying other rules. Except the grass is somewhere that no one would see unless they actually read the rules. The rules were the only place that mentioned that grass.
I like the Van Halen brown M&M version. https://www.smithsonianmag.com/arts-culture/why-did-van-halen-demand-concert-venues-remove-brown-mms-from-the-menu-180982570/
Is the page linked in the site anywhere, or just mentioned in the robots.txt file?
Not sure if that is effective at all. Why would a crawler check the robots.txt if it’s programmed to ignore it anyways?