Dev, It is fully covered in this thread. Some of us have seen an immediate update of new pages added. Some have redesigned sites and pages and had them indexed right away. Some have seen nothing at all.
These are the number of pages G is showing: www.sanlucar-de-barrameda Was 13 now up to 39 All mainly cached BEFORE I did the G sitemap and has only just shown up on G. www.costaysierra.com Was 0 still 0 Can't say anything!!! www.costaandsierra.com Was 10-15 now at 42 pages cached The extra pages seem to be result of the G sitemap www.casaelbosque.com All about the same. Ian
This is good news to me as I just converted my blogger blog to WordPress. Now i'll just tell Google all the new URLs and they'll be happy. Now why doesn't Yahoo do this?
Is there something wrong with my code <?xml version="1.0" encoding="UTF-8"?> <urlset xmlns=http://www.google.com/schemas/sitemap/0.84> <url> <loc>http://www.proadultoutsourcing.com/</loc> <priority>1.0</priority> </url> <url> <loc>http://www.proadultoutsourcing.com/partners.html </loc> </url> <url> <loc>http://www.proadultoutsourcing.com/solutions.html</loc> </url> <url> <loc>http://www.proadultoutsourcing.com/contact.html</loc> </url> <url> <loc>http://www.proadultoutsourcing.com/webmaster.html</loc> </url> <url> <loc>http://www.proadultoutsourcing.com/portfolio/tgp2.html</loc> </url> <url> <loc>http://www.proadultoutsourcing.com/portfolio/banners2.html</loc> </url> <url> <loc>http://www.proadultoutsourcing.com/portfolio/banners2.html</loc> </url> <url> <loc>http://www.proadultoutsourcing.com/photoeditor.html</loc> </url> <url> <loc>http://www.proadultoutsourcing.com/portfolio/photos.html</loc> </url> <url> <loc>http://www.proadultoutsourcing.com/webdesigner.html</loc> </url> <url> <loc>http://www.proadultoutsourcing.com/portfolio/paysites.html</loc> </url> <url> <loc>http://www.proadultoutsourcing.com/portfolio/tgp1.html</loc> </url> <url> <loc>http://www.proadultoutsourcing.com/portfolio/mgp.html</loc> </url> <url> <loc>http://www.proadultoutsourcing.com/portfolio/tgpsites.html</loc> </url><url> <loc>http://www.proadultoutsourcing.com/portfolio/avs.html</loc> </url><url> <loc>http://www.proadultoutsourcing.com/portfolio/fhg.html</loc> </url> <url> <loc>http://www.proadultoutsourcing.com/portfolio/freesites.html</loc> </url> <url> <loc>http://www.proadultoutsourcing.com/videoeditor.html</loc> </url> <url> <loc>http://www.proadultoutsourcing.com/webdeveloper.html</loc> </url><url> <loc>http://www.proadultoutsourcing.com/seo.html </loc> </url> <url> <loc>http://www.proadultoutsourcing.com/condition.html</loc> </url> <url> <loc>http://www.proadultoutsourcing.com/Policy.html</loc> </url> <url> <loc>http://www.proadultoutsourcing.com/outsourcing-solutions/site-map.html</loc> </url> </urlset> this the error An invalid character was found in text content. Error processing resource 'http://www.proadultoutsourcing.com/sitemap.xml'.... <urlset xmlns= It just a simple site map made in xml.. Any suggestion can do this the URL :http://www.proadultoutsourcing.com/sitemap.xml
<?xml version="1.0" encoding="UTF-8" ?> - <urlset xmlns="http://www.google.com/schemas/sitemap/0.84" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.google.com/schemas/sitemap/0.84 http://www.google.com/schemas/sitemap/0.84/sitemap.xsd"> - <url> <loc>http://www.aeforums.net</loc> <lastmod>2005-06-08T09:30:34Z</lastmod> <changefreq>changefreq=hourly</changefreq> <priority>priority=1.0</priority> </url> your mising some lines from the header, maybe thats it
Yes, you should run ls -lR > myfile.txt in your local web directory folder, and it will recursively list all your files from there into the myfile.txt file. Then you upload this file, and get a sitemap in return.
I have no idea what ls -lR > myfile.txt is... call me a noob... PS: What about forum and blog posts that are stored server side?
Sitemaps works in getting your pages crawled incredibly faster. I've been waiting for months for my site to get crawled heavily like my other site, but no dice. 5000 products and only about 10-15 pages would get crawled. I set up a cron job to submit my sitemap automatically and bam, the next day I'm getting crawled 1300 pages, now it's an everyday occurence.....sweeeeet.
It only works for listing your actual files, not for database driven sites. If your forum is in VBuletin you can use my script to generate a basic sitemap. for blogs I know of a wordpress plug-in and another for serendipity.
Just a bit of further information. The pages on our site are tricky to reach via linking, they are normally arrived at via user searches, which obviously google cannot do. For the past few months the number of pages indexed has been stuck around the 1,000 mark. Anyway after submitting a sitemap, google proceeded to visit thousands of the pages, and the number of indexed pages has suddenly increased to 4,500 with some of the datacenters showing 11,500. Not a bad increase within a few days!!! Obviously this will not help you if you have a flat site but it seems that the sitemap enables google to go deeper than it would normally.
If anyone was having difficulties generating a decent sitemap, I came across this today and took it for a test drive. Worked very well. http://nodemap.com/software.php Just download the .exe version and then it will take care of the rest...
Anyone want to help out with a phpbb one? Willing to barter some links in return for it. They already have site map mods so I'm assuming it would just take some editting to make it a xml file.