Duplicate content

Discussion in 'Search Engine Optimization' started by Cyclops, Jan 28, 2005.

  1. #1
    I have seen a lot of discussion on this topic.....there is a view that Google is currently penalising sites that have duplicate content on there pages.

    My question is.....what constitutes duplicate content.

    There is a lot of advice about various products out there.
    It is a reasonable assumption that a lot of sites carry the same articles. Health related articles come to mind straight away.
    Several sites could have the same article from a source in relation to symptoms and strategies to deal with....acne or mumps spring to mind.

    Recipes are another area where duplicate content could apply...moms apple pie recipe could be on several sites.
    There are millions of examples where this can quite legitimately happen.
    Is this duplicate content ?.

    How does Google decide what is and what isn't duplicate content ?.
    Where is the line drawn ?.

    Cyclops.
     
    Cyclops, Jan 28, 2005 IP
  2. joseph29

    joseph29 Peon

    Messages:
    39
    Likes Received:
    0
    Best Answers:
    0
    Trophy Points:
    0
    #2
    I've been asking this question for months ... I hope someone out there can shead some light on this...
     
    joseph29, Jan 28, 2005 IP
  3. TechEvangelist

    TechEvangelist Guest

    Messages:
    919
    Likes Received:
    140
    Best Answers:
    0
    Trophy Points:
    133
    #3
    This is sometimes a big gray area. I'm sure you will see a lot of opinions about this issue. Here's my viewpoint.

    Identical pages (code and content) under either the same or different domains are a clear candidate for a duplicate content penalty.

    Syndicated articles appear to bypass the penalty, which leads me to believe that if you frame the content in a different code structure, you probably won't see the duplicate page penalized.

    I sometimes post similar articles on multiple Web sites, but I alter the content on each so that they are different. I think this adds a layer of protection.
     
    TechEvangelist, Jan 28, 2005 IP
  4. DangerMouse

    DangerMouse Peon

    Messages:
    275
    Likes Received:
    11
    Best Answers:
    0
    Trophy Points:
    0
    #4
    DangerMouse, Jan 28, 2005 IP
  5. chachi

    chachi The other Jason

    Messages:
    1,600
    Likes Received:
    57
    Best Answers:
    0
    Trophy Points:
    0
    #5
    we have one site that appears to have been hit with this filter. Placing &filter=0 at the end of a google search query reveals the site at position #7 for a certain term....without it, it is nowhere to be seen anymore (used to rank for a number of terms, now it ranks for none).

    I noticed yesterday that Gbot was grabbing pages without the www. in front of the domain name as well as with our secure CName (secure.domain. com). It was grabbing regular pages, but putting secure. in front or dropping the www and doing the same. I am hoping this is why they have hit us with some kind of penalty.

    Yesterday, we placed a 301 redirect to send non www. requests to www and excluded any bot from grabbing pages with the secure cname. Hopefully, this will clear things up soon.

    Last year we were hit with a similar penalty as we had 2 domains pointing to the same content (we had forgotten about) and eventually G penalized it. After 301'ing one to the other, the problem cleared up, but took a number of months to get back to "normal". Hopefully, this will not take as long.
     
    chachi, Jan 28, 2005 IP
  6. jamesorr

    jamesorr Peon

    Messages:
    44
    Likes Received:
    0
    Best Answers:
    0
    Trophy Points:
    0
    #6
    We share a lot of content that we write ourselves between sites that we run ourselves. Is there a way to tell if you have been penalized already?
     
    jamesorr, Jan 29, 2005 IP
  7. hex1848

    hex1848 Peon

    Messages:
    62
    Likes Received:
    4
    Best Answers:
    0
    Trophy Points:
    0
    #7
    I was checking a couple of my keywords this morning and noticed this site:

    www.odlp.com/go.php?id=95

    Which I guess I was added to at some point during a link building stent. Did a Live HTTP Headers (firefox) and sure enough, its a 302 redirect.

    Would this cause a duplicate content penalty? The title and everything is identical to my site.

    What could be done to take care of this? I guess I could simply have them remove me from the directory. Any thing else?
     
    hex1848, Jan 31, 2005 IP
  8. candysmith

    candysmith trying not to be evil

    Messages:
    227
    Likes Received:
    13
    Best Answers:
    0
    Trophy Points:
    88
    #8
    Dangermouse thanks for that link, its great!
    at what % similarity do you think it becomes risky?
     
    candysmith, Apr 8, 2005 IP
  9. Will.Spencer

    Will.Spencer NetBuilder

    Messages:
    14,789
    Likes Received:
    1,040
    Best Answers:
    0
    Trophy Points:
    375
    #9
    I have done extensive testing with duplicate content and have come up with very little in the way of useful results. :eek:

    What seems to be the case is this:

    1. Google can detect duplicate content down to the sentence level and penalize for it.

    2. The choice of which site gets penalized appears somewhat random. :eek:

    3. MSN and Yahoo are significantly less savvy than Google at detecting and supressing duplicate content.

    Right now I have two web sites which swap about 20 SERPS a day. Each day the first site goes from page one to nowhere on about 20 terms and the second site goes from nowhere to page one for those terms. Similarly, the second site goes from page one to nowhere for a different set of terms and the first site goes from nowhere to page one for those terms.

    This is not the behavior I was seeing a few months ago, when the first site lost all SERPS, the second site won all SERPS, and the results didn't change until I redesigned both sites and started the war all over again.

    Very occassionally, both sites will show up on page one.

    It's a mystery.
     
    Will.Spencer, Apr 8, 2005 IP
  10. ScrapbookPal

    ScrapbookPal Peon

    Messages:
    38
    Likes Received:
    1
    Best Answers:
    0
    Trophy Points:
    0
    #10
    I had one site that is now a redirect to our current page take the top rankings for a day while the other dropped to outer darkness. Its all back to normal, but kind of weird.
     
    ScrapbookPal, Dec 14, 2005 IP
  11. redQueen

    redQueen Peon

    Messages:
    667
    Likes Received:
    47
    Best Answers:
    0
    Trophy Points:
    0
    #11
    Will.Spencer, I'm just curious about the test you did, how much of a duplicate are your test subjects to each other?

    In my opinion, the SEs as of present are only capable of restricting the same content over and over again for a certain search. It's not a penalty but more like a filter. They don't de-index a site because it has the same articles with another site. I think SEs understand articles and how it works for a site.

    They'd pick site/s that they feel are important and then that's the only things they list in serps. Sometimes, your page would appear, sometimes it wont. It's not a penalty but would feel like it.

    And there's a certain percentage to how similar a page can get to another page (code and content wise). If a page is the exact duplicate of another, then that's something else.

    And i'm guessing it would be in Google's best of interest if we all still believe that there is a duplicate content penalty.

    Of course, this is just from my viewpoint and other people's opinion will vary and i could be totally wrong.
     
    redQueen, Dec 15, 2005 IP
  12. Will.Spencer

    Will.Spencer NetBuilder

    Messages:
    14,789
    Likes Received:
    1,040
    Best Answers:
    0
    Trophy Points:
    375
    #12
    Mine are fairly duplicate. The differences are:

    1. On one site 15 forum post subjects are done server-side in PHP, so they are seen by Googlebot. On the other site the post subjects are displayed client-side by JavaScript, so they are not seen by Googlebot.

    2. One site is displaying 5 co-op ads. The other site displays none.

    3. Both sites run a server-side PERL script which displays one random link on each page, so no two pages will ever be the same.

    4. One site is frequently updated. The mirror site is only mirrored occassionally, so the amount of difference varies based upon how long it has been since I've allowed the mirror to update.

    From my experience I am led to believe that Google's duplicate content filter is amazingly intelligent. However, I then look at the spammers at Answer.com and their cheesy mirroring of Wikipedia and I am led to the exact opposite conclusion.

    In summary: I have no idea. :eek:
     
    Will.Spencer, Dec 15, 2005 IP
  13. dinodonk

    dinodonk Peon

    Messages:
    120
    Likes Received:
    0
    Best Answers:
    0
    Trophy Points:
    0
    #13
    is this dangerous? just put same content with your competitors and *bump* they gone on google
     
    dinodonk, Dec 23, 2005 IP
  14. Will.Spencer

    Will.Spencer NetBuilder

    Messages:
    14,789
    Likes Received:
    1,040
    Best Answers:
    0
    Trophy Points:
    375
    #14
    You could be sued for copyright infringement -- but copyright laws are effectively impossible to enforce unless the copyrighted material is worth millions of dollars.
     
    Will.Spencer, Dec 25, 2005 IP