probably but I recommend that you first remove session ID's. Do you do an amazon call for each page or all the data is in a database?
I did the simple version, just the sitemap.txt with one URL per line. Google hit it in about 2 hrs. Then with 8 hrs every single page in that file was hit. I am currently working on a script to automatically generate a sitemap.xml for google and a sitemap.php for everyone else. Seems silly to go to all the hassle of selecting which files you want in the sitemap and which ones you don't for only one purpose.
Well, interesting... Yahoo! Slurp raped my site today. A site I put online a week or two ago, and I ONLY submitted via normal Google submit site form and 24 hours ago I did the sitemap thingy. Interestingly, GoogleBot did check my sitemap this evening again, about 24 hours after my first submission. No Google rapes yet though.
Interesting possibility for those of us who use Blogger. It looks like the Atom RSS feed can be submitted...Which solves the update issue.
This is great - submitted two simple txt lists, I manually generated all 3,000 forum archive URLs that look like this: www.site.com/forums/archives/index.php/t-1.html Using MS EXCEL. I cut and pasted from the spreadsheet into Dreamweaver and used find and replace to quickly put it in the right format! It was very quick and easy...They're downloaded, lets see what happens!
For those of us with big sites, I'm looking at a few linkchecker type scripts. The idea is to simply have the script spew a list of all url's in your site into a text file from which you can just create the sitemap. If anyone knows of a script that can do this 'out of the box' then it could be useful. At the moment, I'm trying to pursuade Gbot to crawl a new 30K page site - it'll be an interesting test for this sitemap stuff. I'd normally expect Gbot to take a good couple of months to crawl the site.
Just finished setting up an online interface that will generate a Google xml sitemap from a ls -lR directory listing, handy to create a sitemap from your local file system, and then upload that to your web server. Create Google xml sitemap from directory listing
i used that tool but dont see anything besides this <?xml version="1.0" encoding="UTF-8"?> <urlset xmlns="http://www.google.com/schemas/sitemap/0.84"> </urlset>
I also had that error on the script : <?xml version="1.0" encoding="UTF-8"?> <urlset xmlns="http://www.google.com/schemas/sitemap/0.84"> </urlset> The site I submitted in the end was www.ukmp3.co.uk This site had 600 pages until May 2005 when more products were added in, making 1,800 pages. There was also a test shopping area added ( affiliate driven ) which left us with over 50,000 pages. We used Xenu to grab all the hyperlinks and then AddrView to extract the urls in to a .txt file ( 16MB ). This was "fed" to Google. The file was loaded in to Google after about 2 hours and processed fine. Looking at the pages present in Google this has already jumped from 800 to 9,100 so all seems to be working. I was wondering what was stopping someone taking a sitemap, filling it full of 404 pages ( which Google wont like ) and submitting it ! There are many scripts on Hotscripts.com in php for site mapping. best one so far is : http://www.hotscripts.com/Detailed/32948.html and its free. If there was a way of making the site map and creating the xml that would be great. Maybe Google should make one if they want webmaster to go this way .....
I've submitted my site map ... but hey errors occurred.. PARSING ERRORS!!!! Ive read the errors.. and followed some of it that cause the error.. But nothing happen
boohlick , You have to send and xml file that has to be placed on the root of the domain. You failed on both: its a html sitemap and is not in the root. Also, the name has to be "sitemap", not "site-map". Your sitemap has look similar to as it follows: <?xml version=“1.0†encoding=“UTF-8â€?> <urlset xmlns=“http://www.google.com/schemas/sitemap/0.84â€> <url> <loc>WWW.PUT-YOUR-MAIN-URL-HERE.COM</loc> <priority>1.0</priority> </url> <!-- delete this comment: repeat the 3 following lines but changing the url for every url of your site --> <url> <loc>ANOTHER URL </loc> </url> </urlset> Code (markup): Now do manuallly or code a script that does the sitemap for you, but this is how it has to be. For the techies reading this: yes, you can specify many other things, change the priority, etc... but this would ok.
Actually those rules are only formalities. You can have http://www.domain.com/your/site/map/here.xml.gz if you wanted. (It does have to be XML) Google will get it anyway.
hi all, sorry if this has been covered before. what benefit does submitting my sitemap xml to google have if i know for a fact taht all of my pages are crawled and are being indexed by google after they are posted?