About Rejecting Particular Bots
The robots.txt file relies on the goodwill of the bots. If you have some undesirable bots still accessing your website, you can block their requests based on the User Agent. This is how to do that in nginx configuration:
map $http_user_agent $is_undesirable_bot {
default 0;
~evil-spider 1;
~evil-crawler 1;
~evil-bot 1;
}
server {
# ...
if ($is_undesirable_bot) {
return 403;
}
}
And here's how you can block requests from undesirable bots in .htaccess
file with Apache:
RewriteEngine On
RewriteBase /
RewriteCond %{HTTP_USER_AGENT} (evil-spider|evil-crawler|evil-bot) [NC]
RewriteRule (.*) - [F,L]
# ...
Also by me
Django Paddle Subscriptions app
For Django-based SaaS projects.
Django App for You
Django GDPR Cookie Consent app
For Django websites that use cookies.
Django App for You