hello! I plan on using a robots.txt file to block spiders from crawling certain pages. just a quick question since I'm new to this: User-agent: * Disallow: /register Disallow: /login Code (markup): from that code above, I just wanted clarification whether it will block those 2 URL's from being crawled or does the code above mean that it would that block spiders from crawling the register and login directories of my server? thanks.
yes you are right..some spammy bots don't follow the robots.txt instruction..in this case you can track their IP addresses and block them using htaccess..