Hi, wanted to check if this is the right way i should have my robots.txt file for my site http://www.realpracticesolutions.com # robots.txt User-agent: * Disallow: Disallow: /cgi-bin/ Sitemap: http://realpracticesolutions.com/google_sitemap.xml.gz Thanks for your help
You can use like this: User-agent: * Allow: / Disallow: /cgi-bin/ And disallow any files whichever you want to by using word Disallow:
You could simply write it like User-agent: * Disallow: /cgi-bin/ Sitemap: http://realpracticesolutions.com/google_sitemap.xml.gz All other pages/directories will be allowed by default.
The text in red means that you blocked crawlers from accessing any of your sub-directory or even your root directory, change it to Allow. The correct format should be # robots.txt User-agent: * Allow: / Disallow: /cgi-bin/ Sitemap: http://realpracticesolutions.com/google_sitemap.xml.gz You might also want to change google_sitemap.xml.gz to just sitemap.xml.gz just for simplicity reasons