Currently we have a sitemap we provided to Google, in that sitemap for example we have 100,000 url links but google only index and crawl 50,000 why is it? Having additional pages on top of the 100,000 is that a plus? Let me know what you guys think, any advice or suggestions would be appreciated.
that's usually normal.. try build links on your other inner pages, it helps bots crawl and index your other pages..
Google has capacity to indexed all of them, just make sure there are no duplicate contents and meta tags that restrain Google from indexing them all. And Google normally don't show the exact numbers of the indexed pages, only part of it..
Google never says they will index every single page of the web. They index only they think is right. Also, the page that has server error or got blocked by robots.txt etc will not be indexed
1. Generate an xml sitemap for all your 100,000 pages 2. Update your sitemaps at google webmaster tools 3. make sure the sitemap is live on the google WM and its showing the gree tick sighn 4. improve your website posting frequency or updating frequency 5. create and upload a robot file on your FTP and set crawl frequency on webmaster tools 6. internetwork your website and inner page with links these should help you to improve your indexing chances. After all bots also become tired and need some rest so it could take some time as well. cheers and god bless Josh