Logo Questions Linux Laravel Mysql Ubuntu Git Menu
 

Googlebots Ignoring robots.txt? [closed]

I have a site with the following robots.txt in the root:

User-agent: *
Disabled: /

User-agent: Googlebot 
Disabled: /

User-agent: Googlebot-Image
Disallow: /

And pages within this site are getting scanned by Googlebots all day long. Is there something wrong with my file or with Google?

like image 960
Tim Scott Avatar asked Dec 05 '08 18:12

Tim Scott


People also ask

Can robots txt be ignored?

txt file are only instructions. Malicious crawlers will ignore your robots. txt file and crawl any part of your site that is public, so disallow should not be used in place of robust security measures.

How do I fix robots txt error?

Luckily, there's a simple fix for this error. All you have to do is update your robots. txt file (example.com/robots.txt) and allow Googlebot (and others) to crawl your pages. You can test these changes using the Robots.

Can Google crawl without robots txt?

txt file is not required for a website. If a bot comes to your website and it doesn't have one, it will just crawl your website and index pages as it normally would. A robot. txt file is only needed if you want to have more control over what is being crawled.

How do I bypass robots txt disallow?

If you don't want your crawler to respect robots. txt then just write it so it doesn't. You might be using a library that respects robots. txt automatically, if so then you will have to disable that (which will usually be an option you pass to the library when you call it).


1 Answers

It should be Disallow:, not Disabled:.

like image 81
Sean Carpenter Avatar answered Sep 20 '22 13:09

Sean Carpenter