I have a wordpress blog with over 1million posts. What is the best way to arrange my sitemap(s) for proper indexing?
See the Using Sitemap index files (to group multiple sitemap files) section on http://sitemaps.org/protocol.php page
Single sitemap cannot contain more 53000 links and cannot be bigger then 10MB, however the good part is that Google accepts gz compressed sitemaps.
I've pretty much given up on creating a sitemap for the time being. There are ways to link sitemaps together if it's above 50,000 links, Google talks about it somewhere I think, the only problem is getting it. For me, this may end up being a once a month process.
GSiteCrawler has an option to add more spiders at one time, just found that out today, maybe that will help.
Forget normal Sitemap Generators. My Website has 4.2 Mio Pages. I got serversite tool (selfmade), but it takes like 7h to generarte 1,2 M Pages sitemap. I can san like 200 pages each second but it still takes long and my server cpu usage is high, so i can only run it in the night. I have 8 Core with 16 GB ram
gSiteCrawler will create the multiple files (xml) needed by Google. There could be 250 for your site though?