My sitemap contains around 946 Urls, but only 362 of them are indexed in google (according to Google Webmaster Tools). Can anyone tell me what should I do to crawl all 946 Urls in google ?
u can break site map in small part of 100 link/ sitemap. sitemap1, sitemap2, sitemap3...................sitemap10 then in webmaster tool submit all 10 sitemap
they won't index every page. Unless every single page on your site is important (according to google)
Making several Sitemap file is only good and is a requisite from google after 50000 URL. with 1000 URL, there's no need to divide. There are a certain number of reasons why your URL are not being indexed after a long period of time. Have you check the reasons given in the webmaster area? For example if your URL is giving a redirect return code (0302) when opened, it won't be indexed by Google. Try to generate your sitemap through http://www.webrickco.com/buildsitemap.php, it will give you all return codes for each URL after submission by clicking on a line. Relevance is also a factor of indexation. If many of your pages look similar, it won't index them. Based on what criteria? Only they know. I have also not all my pages indexed by google and apparently, there's no obvious reason for that.
As somebody already mentioned here, Google won't index every page. AFAIK Google index all pages directly linked from the home page, while deeper pages are not always indexed. If such pages are linked from external pages, it increases the probability that they will be indexed by Google. Cheers, Sty
It is not really needed until you have more than 50000 URLs. It is not necessary that Google indexes all pages at once. It may take few days or several months. Kailash
google is not gonna index every page!!! some page link image url ... id no text no information for google....
Check your sitmaps' URLs (spell check) if ok then wait because Google bot have a lot of urls in daily bases. its take time
i think it makes a difference if your site is proceeded with www or not. you might have different urls throughout your site, and google is only picking up one or the other, if that makes sense. eg, you might have a page *ttp://yoursite....., and another *ttp://www.yoursite..... Im not sure exactly what this means, but i think it has an effect on the sites being indexed. trouble is, you have a lot of pages to go through and change, it's probably not worth it.
You should always make sure not to have duplicate URLs. Mixing www with non-www, having port numbers in URLs etc. is all bad. Decide to use one (e.g. non-www or www) Beyond that you will need some patience for Google to index all URLs. It's also always an advantage if you can get backlinks to your content.
It takes time... It could be your indexes to old articles/contents of your has been cut by google and replaced with your news ones. Please consider revising your articles once in a while.
Just my two cents... Google was indexing my site pretty well until I submitted a sitemap. Then, not only did my indexing stop - also my traffic died off almost completely - all my rankings tanked. No idea why. My opinion is that Google can find your pages just fine without a sitemap as long as you have a logical structure to your site. I would build links to your top and sub level categories using social bookmarking etc. Also, on page links (in modules etc) to your lower level categories will help alot.