I've dialed it down to only blocking bots that try to impersonate web browsers. If you set a custom User-Agent header, you won't be blocked (unless you're downloading too much). If you try to disguise your traffic as human traffic, you will be blocked.
You may be blocked manually if you try to download too much. "Too much" is not a hard line, it depends on what you're doing and how much spare bandwidth the site has at the time. Basically, if the site is feeling slow, that's when I start going down the list of the top downloaders and blocking people. If you're downloading less than 50GB per day, then I probably won't take notice of you. If you're downloading more than 50GB per day, that's when I start looking at what you're doing and potentially blocking you if it doesn't seem reasonable.
You should set your User-Agent to something containing your bot's name and/or your name or contact info. You're less likely to get blocked if I can check your bot's code on Github or if there's some way to contact you. You're more likely to get blocked if I can't tell what you're doing and I have no way of contacting you. If you try to disguise your traffic as a browser or as human traffic, you will be blocked.
Hotlinking is allowed within reason. Things like posting images for friends on social media or on forums or personal sites is allowed. Things like building apps and alternate UIs for Danbooru are allowed, as long as they're not monetized, they don't contain "Danbooru" in their name, they don't harvest user passwords or API keys, and they don't remove the Referer header. Things like building competing sites that leech our bandwidth aren't allowed, especially if they're monetized (i.e. building a hentai site that hotlinks or proxies all our images and surrounds them with ads).
If you're downloading images for AI purposes, get the 720x720 samples instead of the full size original images. The full set of posts is 9.5 TB. Downloading that much data will take too long and use too much bandwidth. It would take nearly a day even if you could download at a full 1Gbps, which you can't. Just writing that much data to a hard drive would take nearly a day, even if you were copying it straight from one drive to another.
For reference, something like 7 TB of traffic per day is from bots and other things that aren't real people browsing the site. About half of that is from search crawlers like Google and Bing, from hotlinks on other sites (mainly Google Images), and from Discord embeds (1.5 TB alone is from Discord, probably mostly from bots endlessly dumping random images in some shitty unread server somewhere). The other half is from downloaders and scrapers. This traffic is harder on the servers because most of it can't be cached. With human traffic, most of it can be cached because you have a large number of people viewing a small number of images. With bot traffic, most of it can't be cached because you have a large number of bots downloading a large number of random images, so it all goes straight to disk and the disks eventually can't keep up.