So I had this insiders tip on Google and the way they rate websites to mark for spam. I wrote an article on it here. And I already have someone disagreeing with what I have said. I realize Google may use spiders and filters to initially check for duplicate content, but from there it would seem they would have to have employees to check them from there. So what I want is your opinion on Google having employees who view sites or do they strictly use filters to check for duplicate content? Thanks in advance for participating in my thread--and please don't be rough on me, I think it's something that needs discussed..lol.
It is true they do indeed have humans check the search results to filter out spam etc Here is the report http://tinyurl.com/39awbe Happy Reading
It is not that hard to find duplicate contents, Google programmers will just write some codes to check for duplicate contents in the Google database. Human eyes can't catch all duplicate contents, and do it fast.
Surely they use filters to check the duplicate contents. But I dont think there is any manual interference, imho.
I guess that most would be done via computer algorithms but some portions may be marked out for manual if the algorithm found it too complicated, computer is still not powerful enough to do everything.
I can't imagine that Google has the resources (in terms of employees) to manually review every single page for duplicate content. What they do have, though, are some of the smartest people around to develop algorithms that can not only identify duplicate content, but parse through the literally millions and millions of pages out there efficiently enough to stay on top of the ever-changing landscape.
That would be a huge task if checking for duplicate is done by human resource. How much time, money and man power are needed to accomplish it?
I am not saying 100% is done by humans, I would say easily 99.99% done by machines with the doubtful 0.01% done by humans
I think most of the work is done by spiders as there are count less pages on the internet. A little filtering may be is done by Google employees.
My guess is that they use tools to flag duplicate content. But for high PR sites they may do a manual verification before penalizing a site!
In my view Google use some software tools in general condition and do manually in extreme condition to check duplicate content..
It's really hard to check duplicate content manually. It needs a software or spiders to check millions of website in a whole world for a short time.
Google employees can find spam sites and report them, just like anyone else can. I don't think they are willing to work like bots, but I am sure they review all reported spam sites manually.
This is an interesting doc, but to be clear - it's got nothing to do with duplicate content (the word "duplicate" doesn't occur in the doc). It seems clear that to deal with duplicate content issues, Google would HAVE to create automated solutions. Relying on human input for this would be completely impractical. Where human input IS used (which relates to the doc above) is to ensure the relevance of the search results returned is as high as possible. This is ultimately the single goal of all search engines - to return the most relevant results for the query. You can only judge relevance using humans, so G has built up a database of pages with human-judged relevance scores. This can then be used to test out alogorithm changes (and get objective results that show whether things improve). This is what the above doc is about. Howard