I noticed after the last update that many of my pages on http://www.financenewstoday.com went Supplemental. It could have been that way before and I just never noticed. I have written or paid content writers for unique written content. I also have a forum, blog, etc. All updated regularly. Have good backlinks and have done some deeplinking. I am in the process of changing the source code from tables to CSS and div tags but that was just done a couple of days ago and I am still not finished. The site is about 8 1/2 months old. Could the Supplemental Results be a sign of the Sandbox effect or a bigger problem? Also, in my robots.txt file I have the following code, but google still seems to be crawling the /mp/ folder and indexing the preview.php page # FinanceNewsToday.com robots.txt User-agent: * Disallow: /cgi-bin/ Disallow: /mp/ Disallow: /preview.php Any insight would be appreciated
watch matt cutts on his video about supplemental results this might help you http://video.google.com/videoplay?d...8&q=matt+cutts+++"supplemental+results"&hl=en
The sandbox has nothing to do with supplemental results or indexing. Your robots txt file has nothing to do with it either. How long have your pages had original content on them? Do any of the pages contain duplicate content i.e. free articles?
I may have 1 or 2 free articles out of over 200 articles. I have about 90 blog entries all are unique, and forum posts are unique. Original content has been written from the start in Jan of this year.
Also, many of the articles are showing an old Cache (4 or 5 months) but Google has been indexing new pages within a few days, they are just going Supplemental.
Thanks for the video link. That helped some. It is displaying when I do site:www.financenewstoday.com search. So maybe it's no big deal.
I've watched this video several times, and to be honest, I don't think Matt Cutts said anything to explain supplemental results.
i'm in the same boat, along with another bunch of people here. There are numerous threads discussing this.
I went to google....did a site:financenewstoday.com search....found a random page listed as supplimental'd (www.financenewstoday.com/preview.php?id=27)....grabbed a chunck of text and went back to google, put quotes around the chuck of text, and got this page back in my search results - www.financenewstoday.com/debt/creditcounseling/article_index.php?pg=ShowArticle&aID=27 (this page isn't supplimental'd)... so in that case, it looks like duplicate content within your own site caused that page to be tossed into supplimental results. so I grabbed another page at random www.financenewstoday.com/preview.php?id=147 (supplimental'd too)....ran the same test (grabbed chuck of text, when back to google, threw quotes around them) and got this page back www.financenewstoday.com/realestate/buying/article_index.php?pg=ShowArticle&aID=147 Looks like you've got 2 urls for each page...and google's tossing one out. ......strange though....when you do the site: command all your supplimental results show up first....on the bottom of the 10th page (900-1000)they show some of the non-supplimental'd pages.... I have a feeling you might have tripped a filter somewhere
webuildpages, I thought the same thing. That is the print preview for each article. That is why I put the command in the Robots.txt file for the bots to disallow preview.php but for some reason they are still being indexed.
your robots txt has an error. here's your robots.txt User-agent: * Disallow: /cgi-bin/ Disallow: /mp/ Disallow: /preview.php robots txt is for disallowing folders....not pages. the /preview.php won't work. You'll have to add <META NAME="ROBOTS" CONTENT="NOINDEX, NOFOLLOW"> to each page.....or put those in thier own file, and then disallow the entire file in the robots.txt file (easier).
I was unaware of that. It shouldn't be too hard to do since all the preview.php pages are dynamic as so - /preview.php?id=230. I should be able to add it to the header of the preview.php page and it should be picked up from there, correct?