Some time ago, for Bytespider we contacted the IP block owner via abuse@ to obtain a solution for the massive resource hogging and noncompliance with the robots.txt file. They were happy to work with us and cull the spider herd.
Prior to that a few months ago the preview for GPTBot crawled us at a staggering pace and we blocked entire regions of their campus owned IP address. Today they crawl us at a nominal pace, they are crawling now as I write this with no issues abound.
We take spiders seriously because Debian takes
Bugs seriously.
Typo perfectionish.
"The advice given above is all good, and just because a new message has appeared it does not mean that a problem has arisen, just that a new gremlin hiding in the hardware has been exposed." - FreewheelinFrank