Interesting news about wikipedia and Google relationship, read... http://news.bbc.co.uk/2/hi/technology/6335793.stm This relationship you can see almost every keyword SERP in Google, wikipedia in frist page!!.
it will be interesting to see how wikipedia attempts to take business and traffic from the search markets.
yeah right, google isnt just a ranking technology, or a search technology, but a crawling technology too. unless you have the capability to crawl as much of the internet as fast as google (every domain ((one page per domain)) on the internet every 20 minutes) you will never be able to compete with google(thats per crawl server at LEAST). Storing that information, google has there own distributed file system A front end system to deal with the ammount of queries it gets (all 6million a second). And lastly a team of brains that can beat the other team of brains at google. PIerce
I would say Yahoo crawlers are way faster than Google's. You can test it with a new website and see which engine crawls faster. Not too long I posted it here I am sorry, couldn't resist. Lets stay on topic.
search marketing is being more intresting day by day . i think the best think wait and see what will happen
Well, the thing is. Google is run on pure technical knowhow. If any entity could provide relevant search results as relevant as google does when somebody types in a search phrase, they would get a piece of the market. Wiki is run on a business plan. Anybody can program a wiki, with the right experience and technical staff. Not anybody can write an efficient search engine like google. Microsoft and Yahoo, the other big players in search still have a very low share of the search market. Talk is all good, but when you get down to the b-trees, hashes and heterogenous data structures which fuel data mining, most analysts will be lost in their own spaghetti code.
no way!!! only if you abuse the backlinks, yahoo may come faster...but google is always first to show your sites in serps and able to index deeply, unlike yahoo that stops after 40-50 pages
wikipedia, you can see it in first results in any field any language. may be when you search with your domain name you see it first result.
I usually go to Wikipedia when I am looking for real information anyway. In Google, I usually run into pages upon pages of google adsense sites. Go figure.
There's no reason for Wikipedia to be having financial troubles. If they do go under, it's because of the weight of their own stupidity. If you value idealism over financial sense then that's your problem. The biggest flaw with Wikipedia is MediaWiki which is an inefficient mess of code. They need a new front end on their database to cut back on wasted bandwidth and processing power. I can mirror Wikipedia on a PIII 900 no problem using my own front end. MediaWiki is unusable on that system with 2.7 million pages. I'm currently working on getting the updated 4.5 million pages imported into my system. Once it's done it'll be copied onto a $7 a month GoDaddy account. The live version on my GoDaddy account is still using the old 2.7 million page wikipedia dump. Google is having a field day and I already am getting quite a few hits for a wide range of search terms.
You don't have to handle the edit rate and hit rate of wikipedia. Mediawiki is battle tested. At the moment nothing else in that area is. It doesn't help that the foundation can afford all of 2 coders.
Slow software is slow software. They could handle it better if the software wasn't a lumbering behemoth with unnecessary processing power required to run. Just initializing the MediaWiki class takes several seconds on a PIII 900. That's absurd. And then you have to get the data from the database and parse it using their convoluted text parsing code. A minute or so later you finally get your article. Even simplifying the template that displays the articles would save a lot of bandwidth and processing power. Companies would rather use more servers with more processing power and more bandwidth as a crutch to avoid dealing with a software problem. That's fine. But don't bitch about money then. 5 Employees should be able to take a step back from MediaWiki and develop a more efficient front end to the database. Why should they use the inefficient MediaWiki to display content to the masses who don't need all that editing crap? It'd be trivial to default to a streamlined front end that only reads the wiki database until you click the edit button. You'd then be taken to the MediaWiki version on a seperate set of servers that allows editing, etc. Block search engines from those servers and problem solved.
Most of the people employed by wikimedia are not coders. um: http://meta.wikimedia.org/wiki/Wikimedia_servers Search engine spiders do not put a significant load on the servers.
i was talking about crawling not about indexing. google crawls far more than yahoo, regardless of how fast or how much it indexs. My site, 256MB this month to google, 64MB this month to yahoo. Major difference. And the same goes for every month. Daily thats 2,500 pages, vs 400crawled. Pierce