It works for me too. I tested it on this site https://httpstatus.io/ and it does indeed return error 403. It’s probably some rule in your firewall that’s blocking the HTTPStatus bot or Bot Fight Mode is blocking it. Check the logs and analytics on your account.
@cloonan It loads but returns 403 for most of the search engine bots including Google. That’s the issue here.
@WhiteDemonhia Expect DuckDuckGo, for every other bot it responds 403. I used this tool to diagnose.
If you see Google Search Console data below, it reports 4XX response code for all of the urls it crawls. This directly impacts the indexing and ranking of the entire site.
To do this, send about to 5 requests to your site (from google/bing bots or from httpstatus). Then open two tabs, one for analytics and one for events. Depending on the level of your account, it may take up to 5 minutes for the events to appear. So, if nothing appears, wait 2 minutes and refresh the pages until the events show up. Click on events and it will tell you what triggered the firewall.
Thank you for your response. I did follow your instructions and here is what I got. It seems the managed rules trigger 403. I had the old cloudlflare managed rules deployed, I upgraded and deployed new managed rules. Also, tried disabling the rules all together like below. But nothing seems to fix 403 response even for google bots.
In the Wordpress option, it will depend a lot on your case. Even on my Wordpress sites I didn’t need to check the Optimize For WordPress option. So if you have Wordpress, do an extensive test and then decide if it’s possible to leave it disabled.
In Javascript Detections, leave it enabled, as this helps Bot Fight Mode detect a malicious bot.
Static Resources: For me, works fine with it activated. But test yourself.
In your Managed Rules part of the WAF, you can re-enable OWASP and Managed Rulesets. If you try to send a bot through the httpcodetest site, Cloudflare will block it because it’s a fake bot .
Right, now that you’ve done that, send a bot from Google, Bing or another search engine to crawl your site. Send the real bot, not the one from third-party sites. Go to your webmaster and ask the real bot to scan your site.
It can take hours or even days for the Google/etc bot to crawl your site again.
If the real bot continues to be blocked, add a rule to your Firewall to skip known bots.
Whenever I need to use a skip rule, I always like to mark the requests to generate logs.
After that, set the rule position to Last and save. Now wait 30 seconds and try to send a new request for the real bots to crawl your website. This can take another set of hours or days for the bot to re-crawl. So, be patient.
And Always keep checking the WAF Analytics and Events.
A more aggressive rule would be:
(cf.client.bot) or (http.user_agent contains “Google”) or (http.user_agent contains “Bing”) or (http.user_agent contains “google”) or (http.user_agent contains “bing”)
And just mark to skip the Super Bot Fight Mode and save it with the order Last.
And make sure to activate Cloudflare Managed Rules. Because when theses rules are active, Cloudflare will block fake Google and Bing bots, if a malicious bot using Google/Bing user agent bypass Super Bot Fight Mode. I don’t recommend to use the rule to skip user agent Google/Bing without Cloudflare Managed Rules active, because like I said, it would allow malicious actors to bypass your firewall by spoofing their user agents.