Dear webmasters, I am having a question to clarify. If i have published a new post firstly but some one cut and paste into another place. Google had cached that post firstly my post was cached by google secontly.which one google consider as a copy or duplicate content?
Many website owners and article publishers facing this kind of problem where copied articles and posted on higher authority site getting indexed faster and able to rank them, it is possible on your scenario..
I know Google are now making ways on making ways to see which site posted the content first. Hopefully, it will come out soon.
Placing content firstly then in other sites may not help. Google plans on making ways to check which site posted the content first. But at this time, more trusted and bigger site can use your content and will have better ranking then your site.
Actually, they already have ways in finding out authoritative content. That is why authoritative and branded article sites dominates original content rather than website content scrappers. This however isn't perfect ATM.
That's hard to say. Sometimes Google will identify the duplicate content as the original content, for example, the original content is from a small website, then someone copy the original content to a high PR big website, it's possible that Google will take the latter as original content.
i think big g has sophisticated alghorisms about that but the second copy may rank better and google now does not penalize for duplicated content it decide which copy is better
I've been wondering about this too, that s**ks!! it's like someone steals your hard work, google should do something about this
Can you please tell me what are the possibilities Google algorithms have to find out duplicate content with out index.
When your content is copied and published over some high authority site and it gets cached before caching of your content, then no doubt that Google will give first preference to the content over a high authority website.
I thought all these days that Google looks at the timing. If duplicates found Google will penalize the one who published the content late. But this is really surprising it is still under development. I wonder how people dealt with duplicates all these days,
One thing you can do to prevent this mal practices is that you should submit the link of the web page in social bookmarking sites like digg, delicious and directories like dmoz. So that the chance of being index will increase and Google will give the first preference to your site.
I meant the time it was live online. Publishing time. But may be that is not possible. Google will index when the site has been crawled I think. The they crawl first will think that was published earlier.
Eveything is big smelly BS. Google cannot, because no one cannot tell the rights to the article, only the owner can. The publishing time does not matter, the authority does not matter. It is not 'me contra scraper' game. Reuters, or AP syndicate news, you pay for them you have the rights to publish , you do not pay - you are a bad guy. How Google can tell if you paid copyright or not. The only thing Google can do is to make a profile of probable scraper. If you site is close to the profile, you may be out of the game even with clean hands.
comparison of content with the previously cached page by google bot could easily do the content checking... but copy content is still can not be properly tracked by google.. this is self experienced fact...
This is big question what google means when they say duplicate content. If I will quote some part of article, taken from other website, and then I will provide my comments about it, will that be considered as duplicate? What % of page should be identical in order to be considered duplicate? Since websites have different layouts, text in headers, footers, left and right bars, all that is page content. It will be never 100% duplicate content. Does google really capable to decide if content is duplicate? I doubt and I'm sure they will not be able to detect majority of duplicates. Also how they will decide where is original content and who duplicated it. Lets say I have a website that crawled every hour and someone else created new blog and posted few fresh articles. His website will be crawled may be in a week. If I will take his content, Google will index it on my website first. So who will be duplicate in this case? I think what Matt told was more to scare copy website, but really then will not be able to catch them in most of the cases.