Robots.txt
by
Darrell Kingsley
—
last modified
Mar 13, 2014 02:14 PM
This file restricts access to your site for well-behaved spiders and bots
The biggest gotcha is the user-agent disallow statements are not additive, so if you have some rules for all user agents, and then some different rules for Googlebot, Googlebot will only pay attention to the Googlebot rules. To get all the rules to apply, you have to specify the "all bot" rules again within the Googlebot rules:
User-agent: * Disallow: /my_folder User-Agent: Googlebot Disallow: /my_other_folder
Googlebot only ignores my_other_folder. To get it to ignore both you need:
User-Agent: Googlebot Disallow: /my_other_folder Disallow: /my_folder