Hi, Disallowing main urls is easy: Disallow: /myurl/ But I need to disallow subpages which look like this: /myurl/page=1 /myurl/page=2 /myurl2/page=1 /myurl2/page=2 /something/page=1 ... etc I just want to disallow all pages that have charackters like these in their urls: /page= But typing just this doesn't do the job. Is there some way like this: Disallow: /.../page= ???
This will do it: # Disallow: /*? You might also be interested in disallowing certain extensions like this: Disallow: /*.pdf$ See a thread on it at the Search EngineWatch forum: http://forums.searchenginewatch.com/showthread.php?t=13457 And note that only Google pays attentions to some of these commands.
@RockyMtnhi: Regular expressions are not part of robots.txt standard but it's possible that google supports them. @OP: According to the standard there is no way to disallow indexing of such pages.