Brajeshwar

4-min read

Robots.txt, meta tags; Blogger’s Ninja Tool to control how search engines index your sit

I’ve been doing some bit of my own research on how to weed out non-usable contents and show just the good contents to search engines. I began to use robots.txt sometime back but they are limited to just disallow of some folders (like my wordpress installation folder - “wp”). The other day, I was reading the Robots Exclusion Protocol (REP) on Google’s own Blog, and learnt a lot that was missing from my understanding of how you can take control of Search Engines indexing your site’s content.

Meta Tags

Google have recently introduced a new META tag that will allow us to set when we want the page to be removed from the main Google Web Search results. For instance, if you want to remove a particular page after end of this year, then add the following Meta tag to your page (the date format is RFC 850);

META NAME="GOOGLEBOT" CONTENT="unavailable_after: 31-Dec-2007 24:00:00 GMT

However, the REP Meta tags works only for HTML Pages. Nonetheless, Google gave us an option to control access to other documents - Adobe PDF Files, Video and Audio file and many other types. This thus extends the same flexibility for specifying per-URL tags to all other file type. You’ve to simply add any supported Meta tag to a new X-Robots-Tag directive in the HTTP Header used to serve the file.

Here are some examples;

Don’t display a cache link or snippet for this item in the Google search results X-Robots-Tag: noarchive, nosnippet

Don’t include this document in the Google search results X-Robots-Tag: noindex

Tell us that a document will be unavailable after 31st Dec 2007, 12:00 pm GMT X-Robots-Tag: unavailable_after: 31 Dec 2007 24:00:00 GMT

You can combine multiple directives in the same document.

Do not show a cached link for this document, and remove it from the index after 31st Dec 2007, 12:00 pm GMT X-Robots-Tag: noarchive X-Robots-Tag: unavailable_after: 31 Dec 2007 24:00:00 GMT

Robots.txt

Robots.txt allows you to control how search engines access your web site. It allows us to control access at multiple levels – the entire site, through individual directories, pages of a specific type, down to individual pages.

Googlebot specific robots.txt

Google Robots, unlike other bots, allow the use of wildcards - * - to match a sequence of characters. This way, we can do complex Allow and Disallow directive to the Googlebots.

To block all wordpress files from being crawled by Googlebots, we can have

User-Agent: Googlebot
Disallow: /wp-*.php

It can be even in the form of a folder patten - here, a pattern of myimages_xyz can be blocked (where xyz represents your numbers folders or something similar)

User-Agent: Googlebot
Disallow: /myimages_*/
Disallow: /porn/*.jpg

The Googlebot also has an allow tag to allow your files, folders to be crawled by it. This is particularly useful when used in combination with the Wildcard pattern matching scheme to create more complex robots.txt. Here, let’s block a sub-folder on a site but allow some specific folders or files within that sub-folder. Let’s assume that we have installed Wordpress inside a folder called “wp” at the root of the website. So, let’s block the wp folder but allow the /wp/wp-content/uploads/ to be crawled.

User-Agent: Googlebot
Disallow: /wp/
Allow: /wp/wp-content/uploads/

You can do a even more complex Disallow/Allow pattern matching. Let say, if “?” indicates a session ID, we might like to exclude all URLs that contain them to avoid duplicate pages for the Googlebot. However, URLs that ends with a “?” may be the page that we want to include. In this scenario, we can block any URL that includes a “?” but not the one that specifically ends in “?”

User-Agent: Googlebot
Allow: /*?$
Disallow: /*?

A combination and permutation of robots.txt and Meta Tags can help you fine-grain control over your site’s content. Together, robots.txt and META tags give you the flexibility to express complex access policies.

REFERENCES

← Prev Next →