Disallowing pages with 2 variables

Discussion in 'robots.txt' started by GoLinks, Feb 9, 2009.

  1. #1
    Hi all,

    I have some sort of code problem in a website which I cannot disclose its url and I need to remove from google results all pages that have a certain word appearing twice in the url. for example:

    sitename.com/jimi/otherphrases/jimi/morephrases

    so I need to disallow and somehow have noindex, nocache, nofollow to all pages that have the word "jimi" appearing twice.

    Anyone has a clue?
     
    GoLinks, Feb 9, 2009 IP
  2. manish.chauhan

    manish.chauhan Well-Known Member

    Messages:
    1,682
    Likes Received:
    35
    Best Answers:
    0
    Trophy Points:
    110
    #2
    sitename.com/jimi/otherphrases/jimi/morephrases

    is it the same path as in your website??

    If it is the path, you can disallow these urls like:

    user-agent: *
    disallow: /jimi/*/jimi/*
     
    manish.chauhan, Feb 9, 2009 IP
  3. GoLinks

    GoLinks Member

    Messages:
    63
    Likes Received:
    2
    Best Answers:
    0
    Trophy Points:
    43
    #3
    thanks Manish, I have 1 more question:

    if I want to remove all urls that has the symbol ~ in them can I just write:

    disallow: *~* ?

    Will it work?
     
    GoLinks, Feb 9, 2009 IP
  4. OKFunds

    OKFunds Peon

    Messages:
    7
    Likes Received:
    0
    Best Answers:
    0
    Trophy Points:
    0
    #4
    Is it bad to have a keyword appear twice in a url?
     
    OKFunds, Feb 15, 2009 IP