Hi guys, I'd like to publish bunch of reprinted articles, but i'd like to forbid these pages for indexing; what would be the best way? thing is - its dupe content and works fine for PPC but don't wana kill my general site because of this; Thanks for any tips Sandy
There is certains meta tag for this purpose. Noindex, nofollow or whatever you like. Put it on pages which you don't want to be indexed and that's all.
u can use robot.txt file but it will be batter if you use nofollow tag to each page which u dont want to index.
I would use a combo of all methods mentioned above. Some search engines don't obey the noindex/nofollow meta attributes.
robots.txt User-agent: * Disallow: /forum As well in the header <meta name="robots" content="noindex,nofollow">
Yeah, that should keep a majority of all search engines out. Spiders/bots obey what they want to obey. Nothing is 100% foolproof.
Set your robots.txt in one of the pages then put the name of the robot ('Googlebot' is the name of Google's bot).
- list it to be not indexed in robots.txt file - If you have XML Sitemap, do not list it in the xml file - wherever you try to hyperlink to this page from within the content, use "nofollow" attribute in the hyperlink
Using .htaccess might be an option, but I personally think it's overkill. If you're concerned about spiders/bots following your links, use scripted links.