You need a file called robots.txt Take a look here: http://www.robotstxt.org/wc/faq.html and here: http://www.robotstxt.org/wc/faq.html#prevent But this work only with legal spiders. Other content stealers doesn't obey the robots.txt
Add the "nofollow" attribute to the links on your website that you want robots to ignore and not crawl.
Ah so the problem is of that nature.. Well, links on other websites can sometimes be controlled, can't they.. If not, then robots.txt is the way to go. Lines like the one below should do the trick:
The best way I think is a combination of robots.txt and meta tags. meta tags like the following in the header section of your pages will work the best. <meta name="googlebot" content="noarchive,noindex,nofollow,nosnippet" /> <meta name="robots" content="noarchive,noindex,nofollow" /> Code (markup):