I have following code in my robots.txt file User-agent: * Disallow: /index.php/*.html$ Allow: / The reason I added wild-card character in the "Disallow" is to block certain links of our site. So, will it work to block the following type of links? http://www.example.com/index.php/images/article-737.html http://www.example.com/index.php/category-9.html
the syntax you have used within "Disallow" is only applicable to Googlebot (Not for other search engine robots). So if you are going to use "*.html$" kind of syntax, you must put User-agent:Googlebot
Yes it should work. Your syntax is correct. "Allow" line is not needed and better if removed. However if possible, I would prefer adding "noindex" metatags to my pages or redirect them. I think blocking the robots should be the last resort.