Right.. with this whole de-indexing fiasco going on I've been doing a bit of reading on duplicate (and similar) content problems folk were experiencing in 2003. Long time ago I know but it's got me thinking about why my site of 255 pages may of (seemingly) settled at 160'ish across most datacenters. I'm interested in this because a lot of my content on one of my sites is substancially similar (for example, if I have a page on Living in Florida I will also have 4 articles in the same section on Living in Florida, covering different aspects but with sort of, obviously, similar content and titles). Anyway here's an example: Say I have a section on living in Florida on my website: www.florida-info.net/living/index.html In this section are listed the following articles on their own seperate pages: Living in Florida - Finding a Home Living in Florida - Where to Shop Beaches in Florida Tips for Living in Florida Now, if at the end of each of these articles also contained the other links like this: Example of the article "Tips for Living in Florida" Do you think there could be any duplicate or 'similar' content issues to cause any de-indexing or filtering on any of the seperate pages, wether it's due to similar TITLES, content or the actual links on each page used again and again on each of the 4 pages. I will love anyone who replies to this, for long time.. 'cos I believe it could be so important to us all! (plus especially me!) Tip: For those that haven't come across the Duplicate Content Analyzer, check it out. I found it earlier and it lets you know if some of your site is similar or duplicate - it crawls your site site-wide. Note I'm only theorizing/pondering why some sites may of lost a chunk of pages. I don't use co-up or any link exchange programs etc. Google may of hit us with a ton of different reasons to de-index and drop our pages (PR for instance, out-going co-op style links) and I'm just wondering if this could be one of them.. boy this is frustrating. I know I should probably wait but nah.. the theories and questions must go on. Ugh! Pete
Does anyone know what G's current duplicate content filter is set to? Used to be 10 words, then it was 8, now someone tells me it's 6. Much lower and it's going to be easy to have duplicate content when it's really not. I mean, how hard is it to have two websites with completely different topics have the same phrases? 6 common words in a row doesn't seem too hard to do.