Hi all, at any one time on our forum we have at least 50 guests all within the IP range of 180.76.5 usually
How can we block access to these?
Info: http://www.sitepoint.com/forums/showthread.php?600926-Anyone-Dealt-with-the-Baiduspider-Bot
Use htaccess.
None of the IPs in that post match the ones that are guesting on our forum?
you can look into a robots.txt file, or if you upgrade to 2.0 it can tell you the name of the bots making it easier to figure out which one you want to block.
theres no plugin that i can use to just ban an ip range?
Oh wait that feature exists already... i'll just use that. thanks
that will not really stop a search engine bot, you will just now get them crawling your forum and all their results will be the banned message, there are better ways to stop search engine bots
LainaaNone of the IPs in that post match the ones that are guesting on our forum?
?? It's all about baidu, which uses tons of different ip's, see http://whois.domaintools.com/180.76.5.1 & the posts refered in my post above shows 1. many are annoyed by the baidu bot, 2. how to stop any unwanted visits from any source. This particular baidu range is 180.76.0.0 - 180.76.255.255, enter 180.76. in .htaccess and problem solved. Good luck.
putting baidu in your robots.txt file seems to work just fine on my forum without touching .htaccess
Lainaus käyttäjältä: Illori - syyskuu 25, 2011, 12:34:39 IP
putting baidu in your robots.txt file seems to work just fine on my forum without touching .htaccess
You must be right - i just checked the guests log and there are about 50 guests on the website from 180.76.5.* ip range - i guess the 'ban' doesnt stop them from accessing the site.
So i just add these lines to my robots txt file?
User-agent: Baiduspider+(+http://www.baidu.com/search/spider.htm)
Disallow: /
Lainaus käyttäjältä: DavidCT - marraskuu 27, 2009, 10:50:54 AP
In my experience Baiduspider obeys robots.txt. You can either deny it personally or globally, it seems to respect * wildcard.
robots.txt
#Baiduspider
User-agent: Baiduspider
Disallow: /
Some bots refuse to obey robots.txt, if they even bother to check it. Those you can block through your htaccess file.
Baidu has never followed robots.txt for me. I was still getting 40 to 50 of the little bandwidth suckers at one time. I just banned them all in .htaccess-
deny from 180.76
Yeah the robots.txt change did nothing !
I will try the htaccess method.
Thanks