Hello.. Please help me to create a robots.txt file for my vBulletin Forum I want all spiders to crawl my site, but want then NOT to crawl and index some of the folders. And, I'd also like to put a link to sitemap file in it.. How do I do it? Thanks
Examples This example allows all robots to visit all files because the wildcard "*" specifies all robots: User-agent: * Disallow: This example keeps all robots out: User-agent: * Disallow: /
The next is an example that tells all crawlers not to enter four directories of a website: User-agent: * Disallow: /cgi-bin/ Disallow: /images/ Disallow: /tmp/ Disallow: /private/ Example that tells a specific crawler not to enter one specific directory: User-agent: BadBot # replace the 'BadBot' with the actual user-agent of the bot Disallow: /private/
Example that tells all crawlers not to enter one specific file: User-agent: * Disallow: /directory/file.html Note that all other files in the specified directory will be processed. Example demonstrating how comments can be used: # Comments appear after the "#" symbol at the start of a line, or after a directive User-agent: * # match all bots Disallow: / # keep them out
Not every search engines and crawlers following this robot.txt exclusion protocol standard. Only the genuine search engines and spiders are following this protocol standard. Preventing the unwanted robots apart from your sites help you to save more bandwidths Configure your robots.txt in good manner. Google, Yahoo, and Bing always look for robot.txt file in every websites. There are some examples I given below 1. Allows web robots to visit all files User-agent:* Disallow: 2. Allows all web robots out User-agent:* Disallow: / 3. By specifying the directory name between two slashes in Disallow, denies the web spiders to access to those specified directory. User-agent:* Disallow: /cgi-bin/ Disallow: /images/
i ave create and uploade it if anyone want to check it so open below link http://www.redesignunit.com/robots.txt