Hi, Can anyone tell me the correct way to block a set few pages on my site from being indexed/spidered by search engine bots.? I would like to do this via .htaccess. if you could show me a example of a .htaccess file that would block the following i can then cut/paste/edit it into my own .htaccess file. details.php contact.php contact_us.php thanks in advance.
ok ... how ... how do i write in robots.txt to ignore the pages details.php contact.php contact_us.php
User-agent: * Disallow: /details.php Disallow: /contact.php Disallow: /contact_us.php please note, that they are case sensitive and if that file is located in the folder, just add the folder before it, like /folder/filetoblock.php
this is true, only large players obeys this robots.txt, but if some of the crawlers are so called spam bots, then just block them via .htaccess
Yes, you'll need to block those spam bots via .htaccess, but most spam bots are not using their own "bot name" on the user-agent field. Usually they send the user-agent of latest versions of Exporer or Firefox. So you'll need to block by IP and not by user-agent name.