I just just observing my google webmaster tool section, and I see something wrong. 247 pages of my website have this error: "URL restricted by robots.txt" What is this?? Is this stopping my website from getting ranked higher on google? How can I fix this crawling error? Thank you so much
You can check to see what's on your robots.txt from your Webmaster Tools. Some how these have obviously been restricted either by mistake or something altogether more fishy. I suggest checking it out straight away. If you feel the robots.txt is wrong, upload another one (search the internet for this).
Maybe it's not an error, if you are using a CMS like Wordpress for example, you don't want certain pages indexed, e.g. admin pages, certain parts of the archive. It doesn't influence rankings of the other pages, it just excludes certain pages from getting indexed. Short version: It doesn't have to be an error.
Here's what the robots.txt is currently at. Do you think you can glance at it and tell me what's wrong? http://www.freebieshark.com/robots.txt Thank you
Those 200+ pages are actual pages to be visited. None of them really are admin pages. Do you think you can glance at it and see what's wrong with it? http://www.freebieshark.com/robots.txt EDIT: Update, I figured it out. Thank you both for helping me out. Much appreciated.