How many servers does Yahoo have

Original hardwareEdit

The original hardware (circa 1998) that was used by Google when it was located at Stanford University included:[40]

  • Sun Microsystems Ultra II with dual 200MHz processors, and 256MB of RAM. This was the main machine for the original Backrub system.
  • 2 × 300MHz dual Pentium II servers donated by Intel, they included 512MB of RAM and 10 × 9GB hard drives between the two. It was on these that the main search ran.
  • F50 IBM RS/6000 donated by IBM, included 4 processors, 512MB of memory and 8 × 9GB hard disk drives.
  • Two additional boxes included 3 × 9GB hard drives and 6 x 4GB hard disk drives respectively (the original storage for Backrub). These were attached to the Sun Ultra II.
  • SDD disk expansion box with another 8 × 9GB hard disk drives donated by IBM.
  • Homemade disk box which contained 10 × 9GB SCSI hard disk drives.

Production hardwareEdit

As of 2014, Google has used a heavily customized version of Debian (Linux). They migrated from a Red Hat-based system incrementally in 2013.[41]

The customization goal is to purchase CPU generations that offer the best performance per dollar, not absolute performance. How this is measured is unclear, but it is likely to incorporate running costs of the entire server, and CPU power consumption could be a significant factor.[42] Servers as of 20092010 consisted of custom-made open-top systems containing two processors (each with several cores[43]), a considerable amount of RAM spread over 8 DIMM slots housing double-height DIMMs, and at least two SATA hard disk drives connected through a non-standard ATX-sized power supply unit.[44] The servers were open top so more servers could fit into a rack. According to CNET and a book by John Hennessy, each server had a novel 12-volt battery to reduce costs and improve power efficiency.[43][45]

According to Google, their global data center operation electrical power ranges between 500 and 681 megawatts.[46][47] The combined processing power of these servers might have reached from 20 to 100 petaflops in 2008.[48]

Network topologyEdit

Details of the Google worldwide private networks are not publicly available, but Google publications[49][50] make references to the "Atlas Top 10" report that ranks Google as the third largest ISP behind Level 3.

In order to run such a large network, with direct connections to as many ISPs as possible at the lowest possible cost, Google has a very open peering policy.[51]

From this site, we can see that the Google network can be accessed from 67 public exchange points and 69 different locations across the world. As of May 2012, Google had 882 Gbit/s of public connectivity (not counting private peering agreements that Google has with the largest ISPs). This public network is used to distribute content to Google users as well as to crawl the internet to build its search indexes. The private side of the network is a secret, but a recent disclosure from Google[52] indicate that they use custom built high-radix switch-routers (with a capacity of 128 × 10 Gigabit Ethernet port) for the wide area network. Running no less than two routers per datacenter (for redundancy) we can conclude that the Google network scales in the terabit per second range (with two fully loaded routers the bi-sectional bandwidth amount to 1,280 Gbit/s).

These custom switch-routers are connected to DWDM devices to interconnect data centers and point of presences (PoP) via dark fiber.

From a datacenter view, the network starts at the rack level, where 19-inch racks are custom-made and contain 40 to 80 servers (20 to 40 1U servers on either side, while new servers are 2U rackmount systems.[53] Each rack has an Ethernet switch). Servers are connected via a 1Gbit/s Ethernet link to the top of rack switch (TOR). TOR switches are then connected to a gigabit cluster switch using multiple gigabit or ten gigabit uplinks.[54] The cluster switches themselves are interconnected and form the datacenter interconnect fabric (most likely using a dragonfly design rather than a classic butterfly or flattened butterfly layout[55]).

From an operation standpoint, when a client computer attempts to connect to Google, several DNS servers resolve www.google.com into multiple IP addresses via Round Robin policy. Furthermore, this acts as the first level of load balancing and directs the client to different Google clusters. A Google cluster has thousands of servers, and once the client has connected to the server additional load balancing is done to send the queries to the least loaded web server. This makes Google one of the largest and most complex content delivery networks.[56]

Google has numerous data centers scattered around the world. At least 12 significant Google data center installations are located in the United States. The largest known centers are located in The Dalles, Oregon; Atlanta, Georgia; Reston, Virginia; Lenoir, North Carolina; and Moncks Corner, South Carolina.[57] In Europe, the largest known centers are in Eemshaven and Groningen in the Netherlands and Mons, Belgium.[57] Google's Oceania Data Center is located in Sydney, Australia.[58]

Data center network topologyEdit

To support fault tolerance, increase the scale of data centers and accommodate low-radix switches, Google has adopted various modified Clos topologies in the past.[59]

Project 02Edit

Google data center in The Dalles, Oregon

One of the largest Google data centers is located in the town of The Dalles, Oregon, on the Columbia River, approximately 80 miles (129km) from Portland. Codenamed "Project 02", the million[60] complex[further explanation needed] was built in 2006 and is approximately the size of two American football fields, with cooling towers four stories high.[61] The site was chosen to take advantage of inexpensive hydroelectric power, and to tap into the region's large surplus of fiber optic cable, a remnant of the dot-com boom. A blueprint of the site appeared in 2008.[62]

Summa papermillEdit

In February 2009, Stora Enso announced that they had sold the Summa paper mill in Hamina, Finland to Google for 40 million Euros.[63][64] Google invested 200 million euros on the site to build a data center and announced additional 150 million euro investment in 2012.[65][66] Google chose this location due to the availability and proximity of renewable energy sources.[67]

Modular container data centersEdit

In 2005,[68] Google was researching a containerized modular data center. Google filed a patent application for this technology in 2003.[69]

Floating data centersEdit

In 2013, the press revealed the existence of Google's floating data centers along the coasts of the states of California (Treasure Island's Building 3) and Maine. The development project was maintained under tight secrecy. The data centers are 250 feet long, 72 feet wide, 16 feet deep. The patent for an in-ocean data center cooling technology was bought by Google in 2009[70][71] (along with a wave-powered ship-based data center patent in 2008[72][73]). Shortly thereafter, Google declared that the two massive and secretly-built infrastructures were merely "interactive learning centers, [...] a space where people can learn about new technology."[74]

Google halted work on the barges in late 2013 and began selling off the barges in 2014.[75][76]