Does Google Use Filters or Employees to check for duplicate content?

Discussion in 'Google' started by MaryMary, Mar 19, 2008.

  1. #1
    So I had this insiders tip on Google and the way they rate websites to mark for spam. I wrote an article on it here. And I already have someone disagreeing with what I have said.

    I realize Google may use spiders and filters to initially check for duplicate content, but from there it would seem they would have to have employees to check them from there.

    So what I want is your opinion on Google having employees who view sites or do they strictly use filters to check for duplicate content?

    Thanks in advance for participating in my thread--and please don't be rough on me, I think it's something that needs discussed..lol.
     
    MaryMary, Mar 19, 2008 IP
    rohitbhisey and boldy like this.
  2. gibex

    gibex Active Member

    Messages:
    1,060
    Likes Received:
    21
    Best Answers:
    0
    Trophy Points:
    95
    #2
    are millions of sites out there! I don't think that check is possible by G employees
     
    gibex, Mar 19, 2008 IP
  3. netrepreneur

    netrepreneur Peon

    Messages:
    14
    Likes Received:
    1
    Best Answers:
    0
    Trophy Points:
    0
    #3
    It is true they do indeed have humans check the search results to filter out spam etc

    Here is the report http://tinyurl.com/39awbe

    Happy Reading :)
     
    netrepreneur, Mar 19, 2008 IP
    MaryMary likes this.
  4. MaryMary

    MaryMary Prominent Member

    Messages:
    1,982
    Likes Received:
    559
    Best Answers:
    0
    Trophy Points:
    310
    #4
    Awesome! That's not where I got the info from but I'm sure it will be a great read!
     
    MaryMary, Mar 19, 2008 IP
  5. cheapez

    cheapez Active Member

    Messages:
    1,123
    Likes Received:
    14
    Best Answers:
    0
    Trophy Points:
    78
    #5
    It is not that hard to find duplicate contents, Google programmers will just write some codes to check for duplicate contents in the Google database. Human eyes can't catch all duplicate contents, and do it fast.
     
    cheapez, Mar 19, 2008 IP
  6. predpriemach

    predpriemach Peon

    Messages:
    25
    Likes Received:
    0
    Best Answers:
    0
    Trophy Points:
    0
    #6
    yes it easy to find it with a filter see how copyscape.com do it :)
     
    predpriemach, Mar 19, 2008 IP
  7. sray

    sray Active Member

    Messages:
    298
    Likes Received:
    29
    Best Answers:
    0
    Trophy Points:
    70
    #7
    Surely they use filters to check the duplicate contents. But I dont think there is any manual interference, imho.
     
    sray, Mar 19, 2008 IP
  8. wisdomtool

    wisdomtool Moderator Staff

    Messages:
    15,825
    Likes Received:
    1,367
    Best Answers:
    1
    Trophy Points:
    455
    #8
    I guess that most would be done via computer algorithms but some portions may be marked out for manual if the algorithm found it too complicated, computer is still not powerful enough to do everything.
     
    wisdomtool, Mar 19, 2008 IP
    boldy likes this.
  9. GeekBouteek

    GeekBouteek Peon

    Messages:
    14
    Likes Received:
    1
    Best Answers:
    0
    Trophy Points:
    0
    #9
    I can't imagine that Google has the resources (in terms of employees) to manually review every single page for duplicate content. What they do have, though, are some of the smartest people around to develop algorithms that can not only identify duplicate content, but parse through the literally millions and millions of pages out there efficiently enough to stay on top of the ever-changing landscape.
     
    GeekBouteek, Mar 19, 2008 IP
  10. poshswinger

    poshswinger Active Member

    Messages:
    2,527
    Likes Received:
    56
    Best Answers:
    0
    Trophy Points:
    90
    #10
    That would be a huge task if checking for duplicate is done by human resource. How much time, money and man power are needed to accomplish it?
     
    poshswinger, Mar 19, 2008 IP
  11. wisdomtool

    wisdomtool Moderator Staff

    Messages:
    15,825
    Likes Received:
    1,367
    Best Answers:
    1
    Trophy Points:
    455
    #11
    I am not saying 100% is done by humans, I would say easily 99.99% done by machines with the doubtful 0.01% done by humans

     
    wisdomtool, Mar 19, 2008 IP
  12. ari dubov

    ari dubov Peon

    Messages:
    2
    Likes Received:
    0
    Best Answers:
    0
    Trophy Points:
    0
    #12
    Do they review the sites over and over again
     
    ari dubov, Mar 19, 2008 IP
  13. angilina

    angilina Notable Member

    Messages:
    7,824
    Likes Received:
    186
    Best Answers:
    0
    Trophy Points:
    260
    #13
    I think most of the work is done by spiders as there are count less pages on the internet. A little filtering may be is done by Google employees.
     
    angilina, Mar 19, 2008 IP
    boldy likes this.
  14. boldy

    boldy Peon

    Messages:
    561
    Likes Received:
    12
    Best Answers:
    0
    Trophy Points:
    0
    #14
    yes at least on few to results google employees will be monitering
     
    boldy, Mar 19, 2008 IP
  15. dailytechnotes

    dailytechnotes Peon

    Messages:
    8
    Likes Received:
    0
    Best Answers:
    0
    Trophy Points:
    0
    #15
    My guess is that they use tools to flag duplicate content. But for high PR sites they may do a manual verification before penalizing a site!
     
    dailytechnotes, Mar 19, 2008 IP
  16. utrustltd

    utrustltd Peon

    Messages:
    3
    Likes Received:
    0
    Best Answers:
    0
    Trophy Points:
    0
    #16
    I think the google must use filters
     
    utrustltd, Mar 19, 2008 IP
  17. rohitbhisey

    rohitbhisey Banned

    Messages:
    706
    Likes Received:
    25
    Best Answers:
    0
    Trophy Points:
    0
    #17
    In my view Google use some software tools in general condition and do manually in extreme condition to check duplicate content..;)
     
    rohitbhisey, Mar 20, 2008 IP
  18. tsukara

    tsukara Banned

    Messages:
    1,030
    Likes Received:
    6
    Best Answers:
    0
    Trophy Points:
    0
    #18
    It's really hard to check duplicate content manually. It needs a software or spiders to check millions of website in a whole world for a short time.
     
    tsukara, Mar 20, 2008 IP
  19. Nestrer

    Nestrer Well-Known Member

    Messages:
    1,663
    Likes Received:
    103
    Best Answers:
    0
    Trophy Points:
    135
    #19
    Google employees can find spam sites and report them, just like anyone else can.

    I don't think they are willing to work like bots, but I am sure they review all reported spam sites manually.
     
    Nestrer, Mar 20, 2008 IP
  20. howard

    howard Well-Known Member

    Messages:
    59
    Likes Received:
    1
    Best Answers:
    0
    Trophy Points:
    123
    #20
    This is an interesting doc, but to be clear - it's got nothing to do with duplicate content (the word "duplicate" doesn't occur in the doc).

    It seems clear that to deal with duplicate content issues, Google would HAVE to create automated solutions. Relying on human input for this would be completely impractical.

    Where human input IS used (which relates to the doc above) is to ensure the relevance of the search results returned is as high as possible. This is ultimately the single goal of all search engines - to return the most relevant results for the query. You can only judge relevance using humans, so G has built up a database of pages with human-judged relevance scores. This can then be used to test out alogorithm changes (and get objective results that show whether things improve).

    This is what the above doc is about.

    Howard
     
    howard, Mar 20, 2008 IP