Thursday, April 19, 2012

How To Disallow GoogleBot From Accessing Your Website

Googlebot is a search bot software of Google, that collects documents and information from the webpage to build up a searchable index for the Google search engine. Googlebot discovers pages by harvesting all of the links on every page it finds. It then follows these links to other web pages. New web pages must be linked to other known pages on the web in order to be crawled and indexed or manually submitted by the website owners.

A problem which webmasters have often concerned with the Googlebot is that, it eats up a massive amount of bandwidth. This cause websites to exceed their bandwidth limit and be taken down temporarily. This is especially troublesome for mirror sites which host many gigabytes of data

If a website owners wish to disallow the Googlebot to collect information on their site, they can do so with the appropriate directives in a robots.txt file, or by adding a special meta tag to the web page.

This issue can be fixed in two ways. You can fix by either appropriate robots.txt file or meta tag

To my viewpoint it's better to choose meta tag. Just add the following meta tag just before < / head >

<meta name="Googlebot" content="nofollow" />

Currently Googlebot only follows HREF links and SRC links

Please comment here if you have any ask on this issue. You are welcome to share our post


Click Here To add Comment