I developed a Java program to scan unlimited URLs for pagerank. The first 3,000 URLS went by at a rate of 5-6 URLs / second (it finished in just over five minutes). Then, it started giving me errors. The program didn't crash, but I could tell that Google was no longer accepting my requests. Is there a way to bypass these limitations? Or should I distribute the program as is, and recommend users to only scan 3,000 URLs at a time?
use proxies - don't see other way around it. In Scrapebox, with 20-30 proxies I can PR check a list of 200k URLs without problems.
Always use Private Proxies over Public ones though. Good Luck with the tool . Let us know how it goes as I'm also looking for such great tool.