Everytime I submit my Sitemap to Google..I get this error We encountered an error while trying to access your Sitemap. Please ensure your Sitemap follows our guidelines and can be accessed at the location you provided and then resubmit. Any help...
Can u give your sitemap's url? I may check what the problem is In addition, chewck your sites <yousitename>.com/robots.txt file
Are u using wordpress or any CMS ..? Please provide us the link to the sitemap and we can find out any problem
If you can share your URL we might be able to give you exact solutions rather than shooting arrows in dark.
I have had this with adsense, I followed the insructions and it still shows as eroor but still get credited if someone clicks a advert
as i expected tthe problem is in ur robots.txt file .. you are disallowwing google through it. Go to robots.txt on ur domain nd remove Disallow:/
Yep, that's a nasty error to make, but an easy one I've made before, always be careful with the disallow.
That's just the default privacy setting problem with your wordpress installation. Go to Settings -> Privacy and allow search engines to crawl your blog. Right now all your pages also have a meta nofollow,noindex.
I have made changes in the wordpress...But in webmaster under Sitemap this is in the robot.txt User-agent: * Disallow: Sitemap: http://celebfickle.com/sitemap.xml.gz What changes should I make? Here is a robot.txt... Can I copy paste the same in my robot.txt file... http://forums.digitalpoint.com/showthread.php?t=564289
It looks all fine now. Need not change all that .. ofcourse, if you want to then copy paste that code in robots.txt in ur wordpress folder
I'm having the same issue; but when I fixed the sitemap, the wordpress settings, uploaded the new robots.txt, Google still says robots.txt is blocking my domain. umm, no I'm not! Any ideas why? Here's what's in the file: User-agent: * Allow: /* Allow: /wp-content/uploads # Google Image User-agent: Googlebot-Image Disallow: Allow: /* # Google AdSense User-agent: Mediapartners-Google* Disallow: Allow: /* # Internet Archiver Wayback Machine User-agent: ia_archiver Disallow: / # digg mirror User-agent: duggmirror Disallow: / Sitemap: [http stripped due to low post count]data-protectiononline.com/sitemap.xml Sitemap: [http stripped due to low post count]data-protectiononline.com/sitemap.xml.gz
Change all Allow: /* to Allow: / .. there is no wildcard character in robots specification (as much as I know)
My solution to allow crawling of the search engines was found in my Privacy settings in wordpress witched it to ALLOW and all is well.
Don't copy that, you'll make things worse. Just let your robots.txt be how it is i.e. User-agent: * Disallow: Sitemap: [url]http://celebfickle.com/sitemap.xml.gz[/url] Code (markup): This is the standard robots.txt lines used by 90%+ sites. If your not too well versed with SEO, just install a seo plugin 'All-in one seo will be good enough'.
Use only: User-agent: * Disallow: Sitemap: http://celebfickle.com/sitemap.xml Don't add .gz extension for sitemap.
Can you justify the answer ...? gz involves decompressing and then reading the xml file for Search Engines .. how come they are more conventional ..?