Hi @melissa-davis, thanks for getting in touch! The forums are only for support regarding the free version of Wordfence, so the following information is appropriate for all versions of the plugin.
You can often check whether genuine bots from these services will observe robots.txt disallow /
type rules as they publicly state that kind of information, along with user-agent and/or IP ranges. There can be issues with SEO rankings amongst other things if you flat-out decide to block certain crawlers, but of course this is at your discretion.
For example, Ahrefs uses a user-agent
that contains the string, “AhrefsBot” amongst other text so you can use wildcards in the Wordfence > Blocking > Custom Pattern section, inputting *AhrefsBot*
in the “Browser User Agent” field. However, as Ahrefs is considered to be observant of robots.txt, my example may be more appropriate for ones that aren’t, or are pretending to be AhrefsBot.
Our process works by serving a Wordfence-branded block page (before any site content is loaded if the Wordfence firewall is optimized) to any site visitors that trigger your custom blocking rule.
For your information, general treatment of crawlers can also be set in the Rate Limiting section of Wordfence > All Options. I set my Rate Limiting Rules to these values to start with:
Rate Limiting Screenshot
- If anyone’s requests exceed – 240 per minute
- If a crawler’s page views exceed – 120 per minute
- If a crawler’s pages not found (404s) exceed – 60 per minute
- If a human’s page views exceed – 120 per minute
- If a human’s pages not found (404s) exceed – 60 per minute
- How long is an IP address blocked when it breaks a rule – 30 minutes
I also always set the rule to Throttle instead of Block. Throttling is generally better than blocking because any good service understands what happened if it is mistakenly blocked and your site isn’t penalized because of it. Make sure and set your Rate Limiting Rules realistically and set the value for how long an IP is blocked to 30 minutes or so.
Thanks,
Peter.