How to block all robots from indexing/crawling all the html pages in my website. I can't block the folders because all those html pages are distributed in so many folders. So I only want all robots NOT to index/crawl all html pages within my website.
Is this correct (to block all robots from crawling all html pages): User-agent: * Disallow: /.html Allow: /wp-admin/admin-ajax.php Sitemap: mydomain/sitemap.xml
There are only three ways to block all bots. 1) password protect your pages, 2) block the bots on your server, or 3) do not put your pages on the public internet. Anything else that 'works' will block honorable robots but will not stop rogue robots. Since you specifically stated ALL, you will need to do #1 #2, or #3 above.