There are chances that Google penalize you on duplicate content. You can prevent this by stopping google from crawling that page which has duplicate content. If you place following code in meta tags then Google will not index that page. Do not confuse your self with nofollow. nofollow tells Search engine not to pass your page rank link which has nofollow attribute. It will be good if other experts give tips on how to tackle google on various issues.
It's better to block the duplicate versions with Robots.txt Also there is no duplicate content "penalty", Google will just choose the strongest document version to rank the highest.
Do both. Add it to your page content and robots.txt The robots.txt is probably an easier option though because on most dynamic sites youll be using an identical header and it can be hard to customize this to individual pages.
It is better to avoid duplicate content at all costs. What is good in having pages with duplicate content? If you show them to Google your site get penalized, if you hide them from Google, you do not get any traffic from Google... What am I missing? Is there a value to having pages with duplicate content?
From a SEO perspective? None at all, in some cases you just want your content read in as many places as possible, so you distribute your content through your network. But you only really want the SE's coming to one point of impact, not 2 or more.
Your site WILL NOT be punished for duplicate content. As for duplicate content within your own site, it's best to eliminate it purely because you are passing link equity to two documents needlessly. Also because the version you want to rank highest may not, such as a bland printer friendly version. Content that's on other sites can have "some" value, more so if it's altered slightly or you have other text on the page. For instance i published the original Google Thesis that's on Stanford.edu and i get a fair bit of traffic from it because my page title is different to theirs and i have a small amount of other text on the page. For example search "seo google thesis" and i'm #1 and #2 beating Stanford.edu with a duplicate of their own material simply because i have "SEO" in my Title and they don't.
This can happen. Consider you have submitted article to directories but you want it to post on your blog too then it is duplicate content. If you don't want to show this page to google then use this meta tag.
thats not true here you can find a great accurate description of duplicate content http://affiliate-marketing-forums.5...5-duplicate-content-s-not-what-you-think.html
That was good article on duplicate content. I will summarize it. Duplicate content is when you have an exact copy of a site. Page for page, file name for file name, image for image, code for code. And exact replica of a page or site.
I'd rather let the other pages that have the content displayed to be indexed (and I presume you talking about blogs), instead of blocking them I would use excerpts of the content which makes things much easier and better.
Thanks to everyone who responded and answered my question. Yes, what you said makes perfect sense. My blog is penalized as I am posting the same articles I submit to ezinearticles directory. I will try using the above approach and see if my blog's toolbar PR will once again turn white with some green (it is gray at the moment).
Yep. That's why I wonder at people getting so stirred up over whether they should post their articles at more than one article directory. Article directories allow people to use the articles for their sites; by putting your articles there, you are allowing for duplicates of your articles to be used all over the internet.
Using NOINDEX in a document by itself instead of paired with a NOFOLLOW will allow you to still maintain diverse internal linking structures and allow all your pages to inherit PR and trust from their parent pages. NOINDEX will just keep those pages out of the index but child pages are still free to be listed as long as they dont have a NOINDEX meta tag as well.