Google Blogoscoped

Forum

Biggest Site?  (View post)

Philipp Lenssen [PersonRank 10]

Friday, February 17, 2006
14 years ago

Who can find the biggest website using Google’s site operator? I’m starting off with 5,940,000 for site:myspace.com.

Ryan Govostes [PersonRank 1]

14 years ago #

site:cnet.com gets 10,400,000
site:livejournal.com gets 16,200,000
site:com.com gets 13,700,000
site:geocities.com gets 13,900,000
site:google.com gets 42,300,000
site:yahoo.com gets 91,400,000
site:wikipedia.org gets 154,000,000

site:.com gets 7,680,000,000 but I guess that's cheating?

Do I win?

Guest [PersonRank 1]

14 years ago #

I'm offering yahoo.com with 126,000,000 pages (google.com/search?q=site%3Ayah ...).

Niraj Sanghvi [PersonRank 10]

14 years ago #

When I followed that link I saw 163,000,000 pages.

Patrick Konsor [PersonRank 1]

14 years ago #

site:wikipedia.org 154,000,000

randfish [PersonRank 1]

14 years ago #

amazon.com – 92.4 million
surprisingly small – irs.gov – 476 thousand

BrianS [PersonRank 7]

14 years ago #

I'm not sure how legal this one is for the competition .... but I get 7,710,000,000 with *.com

google.com/search?q=site%3A*.c ...

Rituel [PersonRank 1]

14 years ago #

site:wikipedia.org 193,000,000 pages on my DataCenter
site:w3.org 146,000,000
site:outer-court.com 102,000

Philipp Lenssen [PersonRank 10]

14 years ago #

> site:.com gets 7,680,000,000 but I guess that's cheating?

Yes :)

Jim Thompson [PersonRank 1]

14 years ago #

When I search with my preferences set to "any language" and no filtering, I get about 663,000,000 from yahoo.com for . (0.18 seconds)

Peter T Davis [PersonRank 0]

14 years ago #

site:amazon.com

Ben Hurkett [PersonRank 0]

14 years ago #

I would have expected more from 'site:www.archive.org'... I suppose they've prevented the 'bots.

But what's more interesting is that Google shows "Results 1 – 2 of about 2,070,000". Bit of a discrepancy there...

(site:archive.org isn't a lot better, either.)

zmarties [PersonRank 10]

14 years ago #

Some other sites of note:

bbc.co.uk 113,000,000 (or 43,700,000 if you restrict it to www.bbc.co.uk)
ebay.com 20,500,000
microsoft.com 35,700,000

and a special mention for

flickr.com 16,800,000, but reduced to just 1 after duplicates are removed!

notbeinguseful [PersonRank 1]

14 years ago #

site:berkeley.edu
122,000,000

surprisingly close, for a school

Kevin [PersonRank 0]

14 years ago #

site:microsoft.com = 296,000,000

sebah [PersonRank 1]

14 years ago #

site:*.wikipedia.org gets "about 193,000,000 "

/pd [PersonRank 10]

14 years ago #

eh ..this is kewl..lets try the borgs only..

typepad = > 14,000,000
livejournal =>9,460,000

wordpress = >1,910,000

msnspaces = >35,500

ok..this kinda forking now... :)-

*.com =>3,960,000,000
*.edu =>1,630,000,000
*.net => 201,000,000
*.org 1,880,000,000

dichotmay of tla..:)-

*.cn =>30,500,000
*.us=>70,100,000
*.in=>8,760,000
*.jp=>150,000,000 <== WTF ???

/pd [PersonRank 10]

14 years ago #

ok I forgot my own country -canada

*ca= >161,000,000
vs
*us=>70,100,000

hmmmmmmmm... poor buggers...big brother is not really big after all :)-

pablotossi [PersonRank 0]

14 years ago #

site:yahoo.com
Resultados 1 – 20 de aproximadamente 664,000,000 de en el dominio yahoo.com. (0.08 segundos)

site:wikipedia.org
Resultados 1 – 20 de aproximadamente 151,000,000 de en el dominio wikipedia.org. (0.33 segundos)

Stephen [PersonRank 0]

14 years ago #

site:bbc.co.uk 158,000,000

Ken [PersonRank 0]

14 years ago #

dmoz.org = 120,000,000

alek [PersonRank 10]

14 years ago #

site:yahoo.com = 544,000,000 ... what do I win?!? ;-)

alek [PersonRank 10]

14 years ago #

And here's a screengrab for 'ya:
  


komar.org/images/2006_02_16_bl ...

Eric Baillargeon [PersonRank 2]

14 years ago #

Alexa ;-)

Support Freedom! [PersonRank 10]

14 years ago #

Google.com site:.com returns 7,680,000,000
Google.cn site:.com returns 3,940,000,000

Three BILLION, Seven hundred forty million pages which Google China censored.

And that's just the "dot coms"

Google.com <html> (to get all html pages) 8,490,000,000

Google.cn <html> (to get all html pages) 4,420,000,000

Thus Four BILLION, seventy million pages which the cowards in Beijing are too terrified to let their citizens see.

I'm stunned.

Andrew Hitchcock [PersonRank 10]

14 years ago #

"Support Freedom!", by your logic, non-BigDaddy servers are censoring 16 billion pages (25 billion versus 9 billion).

Travis Harris [PersonRank 10]

14 years ago #

Results 1 – 10 of about 1,150,000,000 from *.gov for . (0.11 seconds)

(all us Govt so it is only 1 big company.... I think I take the prize!)

Travis Harris [PersonRank 10]

14 years ago #

I know it is not even in the running... but interestingly enough

Results 1 – 10 of about 62,700,000 from mit.edu for . (0.36 seconds)

Travis Harris [PersonRank 10]

14 years ago #

Sick of me yet?
Results 1 – 10 of about 150,000,000 from w3.org for . (0.07 seconds)
(course wikipedia has 193 for me here too... so taht seems to be doing well)

Elias KAI [PersonRank 10]

14 years ago #

site:apple.com 30,900,000 from apple.com for . (0.10 seconds)
site:msn.com 32,000,000 from msn.com for . (0.06 seconds)
site:yahoo.com 103,000,000 from yahoo.com for . (0.07 seconds)
site:google.com 45,800,000 from google.com for . (0.04 seconds)
site:wikipedia.org 155,000,000 from wikipedia.org for . (0.06 seconds)
site:aol.com 14,900,000 from aol.com for . (0.03 seconds)

But as We can See the Fastest from Sweden is AOL.com , the biggest Google.com

Tadeusz Szewczyk [PersonRank 10]

14 years ago #

Enlarge your Index! Google Blogoscoped ;-)

unitedstatians [PersonRank 0]

14 years ago #

site:microsoft.com 42,300,000 from microsoft.com in . (0.17 seconds)
site:apple.com 30,900,000 from apple.com in . (0.10 seconds)

site:slashdot.org 15,000,000 from slashdot.org in . (0.08 seconds)
site:digg.com 5,380,000 from digg.com in . (0.06 seconds)

site:intel.com 2,110,000 from intel.com in . (0.20 seconds)
site:amd.com 699,000 from amd.com in . (0.17 seconds)

site:nvidia.com 182,000 from nvidia.com in. (0.22 seconds)
site:ati.com 78,000 from ati.com in . (0.38 seconds)

site:nasa.gov 46,700,000 from nasa.gov in . (0.40 seconds)
site:esa.int 1,750,000 from esa.int in . (0.18 seconds)
site:nasda.go.jp 98,500 from nasda.go.jp for . (0.22 seconds)

Igor [PersonRank 1]

14 years ago #

site:w3.org gets 546,000,000
site:wikipedia.org gets 155,000,000

When I search with my preferences set to "any language".

Sudeep [PersonRank 1]

14 years ago #

Blogspot gets :

24,100,000 from blogspot.com

AMIT Goyal [PersonRank 1]

14 years ago #

Results 1 – 10 of about 40,600,000 for site: blogspot.com. (0.31 seconds)

FerRory [PersonRank 1]

14 years ago #

*.nl found: 329.000.000
site:.nl 114.000.000

noname [PersonRank 4]

14 years ago #

-site:.com
   about 10,700,000,000 for -site:.com
What did I win ;-)

BTW googlefight is a bit strange:
googlefight.com/index.php?lang ...

Tony Ruscoe [PersonRank 10]

14 years ago #

One of the BigDaddy servers is giving 469,000,000 for site:ebay.com:

66.249.93.104/search?q=site%3A ...

That's much bigger than some of the other servers are returning (i.e. 12,900,000).

/pd [PersonRank 10]

14 years ago #

how accurate are the results ?? I am finding that the results are varying... why is this ?? does this mean the the distrubted DC are funneling 'stale' information out and not in sync with one another in near real time ??

Corsin Camichel [PersonRank 10]

14 years ago #

You know you only can access max. 1000 search results? So you never can check if there are really that much sites ;)

Niraj Sanghvi [PersonRank 10]

14 years ago #

/pd: I remember reading somewhere that the datacenters are always at a different point of indexing, so they always return slightly different results. I would guess there still has to be some level of synchronization so the results aren't wildly different, and so google doesn't crawl the web X 8 (or however many datacenters they use).

/pd [PersonRank 10]

14 years ago #

Niraj: I agree that the indexing at the cache level will always be different. But difference in the 100-200Million pg(s)? is askew. After all I don't think that so0 many pages are created and cached within a 24 hr cycle..or even a 48 hr cycle

or maybe my thought process is flawed and I my brain needs rewiring :)-

pk_synths [PersonRank 2]

14 years ago #

site:tripadvisor.com = 12,400,000

Andi [PersonRank 1]

14 years ago #

Not very big but nevertheless interesting:

site:.eu

google.com/search?hl=de&q= ...

fishfish [PersonRank 0]

14 years ago #

Does the number change much from day to day?

at [PersonRank 0]

14 years ago #

google.com/search?q=site%3A.ar ...

94

Scott [PersonRank 0]

14 years ago #

site:imdb.com = 230,000,000

Scott [PersonRank 0]

14 years ago #

site:mit.edu = 305,000,000

Andi [PersonRank 1]

14 years ago #

site:.eu

20.000

google.com/search?hl=de&q= ...

Let's see how many there'll be tomorrow or so.

Slacter [PersonRank 0]

14 years ago #

Good grief, who maintains all of these pages?!?

SamD [PersonRank 1]

14 years ago #

Does personalised results have an effect on the number?

Or do they just effect the ordering of the same content?

(also: people can remove results now using the link provided, although I don't really know why anyone would do this).

Jürgen [PersonRank 1]

14 years ago #

I don't think this results are really showing the number of pages (and finally it's called "results" and not "pages"). When I search for all results on my own website I get 86,600 results and that is way to much. I never ever have more than 80T pages of my website!

google.com/search?hl=en&q= ...

But there is many dynamic content, maybe this will lead to such a big number?

This thread is locked as it's old... but you can create a new thread in the forum. 

Forum home

Advertisement

 
Blog  |  Forum     more >> Archive | Feed | Google's blogs | About
Advertisement

 

This site unofficially covers Google™ and more with some rights reserved. Join our forum!