My blog has pagination links that allows users and bots alike to access older posts page-by-page. I don't actually want to have those pages indexed, just the posts, so I added: Disallow: /?page to my robots.txt. Unfortunately, it has occurred to me that this may be preventing the bots from accessing the old pages (to keep the older posts indexed) altogether instead of just not indexing them. Is that correct?
the entire robots.txt looks like: User-agent: * Disallow: /admin/ Disallow: /?page The site is http://www.pjhyett.com
Your robots.txt disallows: - http://www.pjhyett.com/?page=2 - http://www.pjhyett.com/?page=14 It does not disallow: - http://www.pjhyett.com/posts/184-what-s-your-anti-code - http://www.pjhyett.com/posts/15-friday-night-s-alright It could be more difficult for the bots to find the pages with your posts if they do not visit the archived pages (/?page=...) though. Jean-Luc