Ive realised it ignores the robots.txt...
User-agent: BaiDuSpider
Disallow: /
User-agent: Baidu
Disallow: /
User-agent: Baiduspider+(+
http://www.baidu.com/search/spider.htm)
Disallow: /
User-agent: Baiduspider+(+
http://www.baidu.com/search/spider_jp.html)
Disallow: /
The only annoying thing is that it visits like every 5 - 10 mins. Its eating my bandwidth.
I cant block it by IP either cause it has loads of IPs