in General

Robots.txt, meta tags; Blogger’s Ninja Tool to control how search engines index your site

Robots

I’ve been doing some bit of my own research on how to weed out non-usable contents and show just the good contents to search engines. I began to use robots.txt sometime back but they are limited to just disallow of some folders (like my wordpress installation folder – “wp”). The other day, I was reading the Robots Exclusion Protocol (REP) on Google’s own Blog, and learnt a lot that was missing from my understanding of how you can take control of Search Engines indexing your site’s content.

Meta Tags

Google have recently introduced a new META tag that will allow us to set when we want the page to be removed from the main Google Web Search results. For instance, if you want to remove a particular page after end of this year, then add the following Meta tag to your page (the date format is RFC 850);

META NAME="GOOGLEBOT" CONTENT="unavailable_after: 31-Dec-2007 24:00:00 GMT

However, the REP Meta tags works only for HTML Pages. Nonetheless, Google gave us an option to control access to other documents – Adobe PDF Files, Video and Audio file and many other types. This thus extends the same flexibility for specifying per-URL tags to all other file type. You’ve to simply add any supported Meta tag to a new X-Robots-Tag directive in the HTTP Header used to serve the file.

Here are some examples;

→ Don’t display a cache link or snippet for this item in the Google search results
X-Robots-Tag: noarchive, nosnippet

→ Don’t include this document in the Google search results
X-Robots-Tag: noindex

→ Tell us that a document will be unavailable after 31st Dec 2007, 12:00 pm GMT
X-Robots-Tag: unavailable_after: 31 Dec 2007 24:00:00 GMT

You can combine multiple directives in the same document.

→ Do not show a cached link for this document, and remove it from the index after 31st Dec 2007, 12:00 pm GMT
X-Robots-Tag: noarchive
X-Robots-Tag: unavailable_after: 31 Dec 2007 24:00:00 GMT

Robots.txt

Robots.txt allows you to control how search engines access your web site. It allows us to control access at multiple levels — the entire site, through individual directories, pages of a specific type, down to individual pages.

Googlebot specific robots.txt

Google Robots, unlike other bots, allow the use of wildcards – * – to match a sequence of characters. This way, we can do complex Allow and Disallow directive to the Googlebots.

To block all wordpress files from being crawled by Googlebots, we can have

User-Agent: Googlebot Disallow: /wp-*.php

It can be even in the form of a folder patten – here, a pattern of myimages_xyz can be blocked (where xyz represents your numbers folders or something similar)

User-Agent: Googlebot Disallow: /myimages_*/ Disallow: /porn/*.jpg (see, I can block google from looking at my porn images ;-) )

The Googlebot also has an allow tag to allow your files, folders to be crawled by it. This is particularly useful when used in combination with the Wildcard pattern matching scheme to create more complex robots.txt. Here, let’s block a sub-folder on a site but allow some specific folders or files within that sub-folder. Let’s assume that we have installed WordPress inside a folder called “wp” at the root of the website. So, let’s block the wp folder but allow the /wp/wp-content/uploads/ to be crawled.

User-Agent: Googlebot Disallow: /wp/ Allow: /wp/wp-content/uploads/

You can do a even more complex Disallow/Allow pattern matching. Let say, if “?” indicates a session ID, we might like to exclude all URLs that contain them to avoid duplicate pages for the Googlebot. However, URLs that ends with a “?” may be the page that we want to include. In this scenario, we can block any URL that includes a “?” but not the one that specifically ends in “?”

User-Agent: Googlebot Allow: /*?$ Disallow: /*?

A combination and permutation of robots.txt and Meta Tags can help you fine-grain control over your site’s content. Together, robots.txt and META tags give you the flexibility to express complex access policies.

REFERENCES

[poll=26]

  1. Hey friend....I was just wondering my adsense account for blogger does not get approved, it's probably as google says it cannot crawl pages and thereforeI dont get access....can u help ??

  2. Hey friend....I was just wondering my adsense account for blogger does not get approved, it's probably as google says it cannot crawl pages and thereforeI dont get access....can u help ??

Comments are closed.

Webmentions

  • Excellent flash resources » Blog Archive » Creating robots.txt files manually Nov 12, 2008

    [...] For additional information on robots.txt files: http://www.brajeshwar.com/2007/robotstxt-meta-tags-bloggers-ninja-tool-to-control-how-search-engines... [...]

  • Robots.txt, meta tags; Blogger’s Ninja Tool to control how search engines index your site « FlashColony webmaster’s blog Nov 12, 2008

    [...] So what is the myster about robots.txt ? Have you ever seen Googlebot browsing your website? Well, it is the way how search engines index your website. More about it here. [...]