Shoemoney's robots.txt with WordPress

Discussion in 'Search Engine Optimization' started by sixfigureblogger, Mar 4, 2008.

  1. #1
    I just read an interesting post on Shoemoney.com and have one question about it. Here's his article:

    http://www.shoemoney.com/2008/03/03/wordpress-robotstxt-tips-against-duplicate-content/

    In his robots.txt file he has it set to disallow index.php like this:

    Disallow: /index.php

    I'm wondering for WordPress if this is a bad idea? I don't like to question a man who brings home da bling, but it does leave me scratching my head a bit.

    Any thoughts about this?
     
    sixfigureblogger, Mar 4, 2008 IP
  2. thor22

    thor22 Peon

    Messages:
    580
    Likes Received:
    8
    Best Answers:
    0
    Trophy Points:
    0
    #2
    I dont see where he sAYS DISallow index.php
     
    thor22, Mar 4, 2008 IP
  3. spdev

    spdev Peon

    Messages:
    65
    Likes Received:
    0
    Best Answers:
    0
    Trophy Points:
    0
    #3
    i see in the robots.txt the disallow for index.php, why is that and how is that beneficial at all? im confused
     
    spdev, Mar 4, 2008 IP
  4. wd_2k6

    wd_2k6 Peon

    Messages:
    1,740
    Likes Received:
    54
    Best Answers:
    0
    Trophy Points:
    0
    #4
    He doesn't say it, it's in his robots.txt file for that website.

    Taking a guess i think it is because on his index.php he is siimply displaying the posts which can also be found on there permanant links therefore there is a duplicate content issue.
     
    wd_2k6, Mar 4, 2008 IP
  5. cableboy

    cableboy Peon

    Messages:
    32
    Likes Received:
    1
    Best Answers:
    0
    Trophy Points:
    0
    #5
    My best says to avoid duplicate content
     
    cableboy, Mar 4, 2008 IP
  6. sixfigureblogger

    sixfigureblogger Banned

    Messages:
    412
    Likes Received:
    7
    Best Answers:
    0
    Trophy Points:
    0
    #6
    I suppose his sitemap will pick up the the blog posts somehow? If he has one?

    I suppose that's the only reasoning behind blocking index.php... to avoid dup content, but I wonder how the posts get indexed if he doesn't have a sitemap?

    :eek:
     
    sixfigureblogger, Mar 4, 2008 IP
  7. nfd2005

    nfd2005 Well-Known Member

    Messages:
    295
    Likes Received:
    20
    Best Answers:
    0
    Trophy Points:
    130
    #7
    I think the thought behind disallowing index.php is because index.php and the root domain is the same thing.

    For example: www.yourdomain.com/ is the same as www.yourdomain.com/index.php. So disallowing index.php will get rid of duplicate content from your root.

    If that makes sense.
     
    nfd2005, Mar 4, 2008 IP
  8. Bohol

    Bohol Peon

    Messages:
    2,680
    Likes Received:
    75
    Best Answers:
    0
    Trophy Points:
    0
    #8
    Can we deploy the blocking on forum sites, too?
     
    Bohol, Mar 4, 2008 IP
  9. sixfigureblogger

    sixfigureblogger Banned

    Messages:
    412
    Likes Received:
    7
    Best Answers:
    0
    Trophy Points:
    0
    #9
    Yeah, I think you're right. I'm taking a wild guess that the ultimate goal is to perhaps only index the blog posts.
     
    sixfigureblogger, Mar 4, 2008 IP
  10. Dan Schulz

    Dan Schulz Peon

    Messages:
    6,032
    Likes Received:
    437
    Best Answers:
    0
    Trophy Points:
    0
    #10
    Actually they're treated as separate pages. THAT is why it's blocked by the robots.txt file.
     
    Dan Schulz, Mar 4, 2008 IP