how to use robot.txt?

Discussion in 'Search Engine Optimization' started by PaulRigby, Nov 28, 2011.

  1. #1
    how to use robot.txt? i dont want to catch the particular page in my website.
     
    PaulRigby, Nov 28, 2011 IP
  2. sophia911

    sophia911 Peon

    Messages:
    241
    Likes Received:
    0
    Best Answers:
    0
    Trophy Points:
    0
    #2
    It stop from Spider catching your page .
     
    sophia911, Nov 28, 2011 IP
  3. sisatel

    sisatel Active Member

    Messages:
    1,391
    Likes Received:
    10
    Best Answers:
    1
    Trophy Points:
    90
    #3
    sisatel, Nov 28, 2011 IP
  4. Black4c3

    Black4c3 Peon

    Messages:
    12
    Likes Received:
    0
    Best Answers:
    0
    Trophy Points:
    0
    #4
    just see on other website...

    you will find what u want.. :)
     
    Black4c3, Nov 29, 2011 IP
  5. danielwood

    danielwood Member

    Messages:
    115
    Likes Received:
    1
    Best Answers:
    0
    Trophy Points:
    28
    #5
    If you don't want SEs to index a particular page of your site, then add the following code in your robots.txt file -

    User-agent: *
    Disallow: /folder/file.html (give the complete path for your file)
     
    danielwood, Nov 29, 2011 IP
  6. leonidasblaze

    leonidasblaze Peon

    Messages:
    602
    Likes Received:
    5
    Best Answers:
    0
    Trophy Points:
    0
    #6
    Robot.txt file has some simple directives which manage the bots.These are define as below.
    1.User agent : The * is a wildcard which means all bots or googlebot for google.
    2.Disallow : Define which folder or file will be executed.
     
    leonidasblaze, Nov 29, 2011 IP
  7. triforce.co.in

    triforce.co.in Peon

    Messages:
    19
    Likes Received:
    0
    Best Answers:
    0
    Trophy Points:
    0
    #7
    [TABLE="width: 658"]
    [TR]
    [TD="width: 658"]http://www.robotstxt.org/robotstxt.html
    Refer this link for using robots.txt file.
    [/TD]
    [/TR]
    [/TABLE]
     
    triforce.co.in, Nov 29, 2011 IP
  8. PaulRigby

    PaulRigby Active Member

    Messages:
    290
    Likes Received:
    1
    Best Answers:
    0
    Trophy Points:
    73
    #8
    how can i see others website robot.txt?
     
    PaulRigby, Nov 30, 2011 IP
  9. jashlok

    jashlok Peon

    Messages:
    125
    Likes Received:
    1
    Best Answers:
    0
    Trophy Points:
    0
    #9
    write the code in notepad and save the notepad in robot.txt

    User-agent: *
    Disallow: /search
    Disallow: /groups
    Disallow: /images
    Disallow: /admin

    If u don't want to crawl the folder and pages of websites then use robots.txt file
     
    jashlok, Nov 30, 2011 IP
  10. orexmedia

    orexmedia Peon

    Messages:
    11
    Likes Received:
    0
    Best Answers:
    0
    Trophy Points:
    0
    #10
    [TABLE="width: 100%"]
    [TR]
    [TD="width: 75%"]A "robots.txt" file is a text file placed on your server which contains a list of robots and "disallows" for those robots. Each disallow will prevent any address that starts with the disallowed string from being accessed.[/TD]
    [/TR]
    [/TABLE]
     
    orexmedia, Nov 30, 2011 IP
  11. marylandlawoffice

    marylandlawoffice Member

    Messages:
    395
    Likes Received:
    2
    Best Answers:
    0
    Trophy Points:
    28
    #11
    write the code in notepad and save notepad Name as robot.txt and put on root path (just like http://www.xyz.com/robot.txt)

    User-agent: *
    Disallow: /search
    Disallow: /groups
    Disallow: /images
    Disallow: /admin
     
    marylandlawoffice, Nov 30, 2011 IP