1. robots.txt blocking bot access to certain locations on your site
2. nginx config that tarpits bots on URLs that they are not supposed to visit
3. some stats on how much resources bots wasted already being tarpited.
2.1 while tarpiting, trickle out, byte by byte, something inane or otherwise poisoning the well.
$i = 65;
$path = mb_convert_case($_SERVER['REQUEST_URI'], MB_CASE_LOWER, "UTF-8");
$end = explode('?w=', $path);
for($i = 65; $i <= 90; $i++)
echo "<a href=\"tarpit.php?w=" . $end . chr($i) ."\">" . chr($i) . "</a><br>";
@timp @rysiek You could just create a page that accepts any characters in the URL path and then make it render a link to the "next" page which is a procedurally generated string of characters based on the provided input. That generates an infinite* chain
* Unless you create a collision in your procedural function; you could even just use a hashing function
This Mastodon instance is for people interested in technology. Discussions aren't limited to technology, because tech folks shouldn't be limited to technology either!