Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Google data centers

From Wikipedia, the free encyclopedia
Facilities containing Google servers

Former Google data center inEemshaven, Netherlands
External videos
YouTube logo
YouTube
video iconGoogle Data Center 360° Tour

Google uses largedata center facilities to providetheir services, which combine large drives, computer nodes organized in aisles of racks, internal and external networking, environmental controls (mainly cooling and humidification control), and operations software (especially as concernsload balancing andfault tolerance).

There is no official data on how manyservers are in Google data centers, butGartner estimated in a July 2016 report that Google at the time had 2.5 million servers. This number is changing as the company expands capacity and refreshes its hardware.[1]

Locations

[edit]

The locations of Google's various data centers by continent are as follows:[2][3]

ContinentLocationGeoProducts LocationCloud LocationTimelineDescription
North AmericaArcola (VA), USA38°56′35.99″N77°31′27.61″W / 38.9433306°N 77.5243361°W /38.9433306; -77.5243361Loudoun CountyN. Virginia (us-east4)2017 - announced[4][5]
North AmericaAtlanta (GA), USA33°44′59.04″N84°35′5.33″W / 33.7497333°N 84.5848139°W /33.7497333; -84.5848139Douglas County-2003 - launched350 employees
South AmericaCerrillos, Santiago, Chile33°31′14″S70°43′18″W / 33.520515°S 70.721695°W /-33.520515; -70.721695[6]-Santiago (southamerica-west1)2020 - announced[7]

2021 - launched[8]

AsiaChanghua County, Taiwan24°08′18.6″N120°25′32.6″E / 24.138500°N 120.425722°E /24.138500; 120.425722Changhua CountyTaiwan

(asia-east1)

2011 - announced

2013 - launched

60 employees
North AmericaClarksville (TN), USA36°37′16″N87°15′47″W / 36.6211599°N 87.2630735°W /36.6211599; -87.2630735Montgomery County-2015 - announced
North AmericaColumbus (OH), USA-Columbus (us-east5)2022 - launched[9]
North AmericaCouncil Bluffs (IA), USA41°13′17.7″N95°51′49.92″W / 41.221583°N 95.8638667°W /41.221583; -95.8638667Council Bluffs2007 - announced

2009 - completed first phase completed

2012 and 2015 - expanded

130 employees
North AmericaCouncil Bluffs (IA), USA41°10′06″N95°47′46″W / 41.1682533°N 95.7961246°W /41.1682533; -95.7961246Iowa (us-central1)
AsiaDelhi, India-Delhi (asia-south2)2020 - announced

2021 - launched[10]

Middle EastDoha, Qatar-Doha (me-central1)2023 - launched[11]
EuropeDublin, Ireland53°19′12.39″N6°26′31.43″W / 53.3201083°N 6.4420639°W /53.3201083; -6.4420639Dublin-2011 - announced

2012 - launched

150 employees[12]
EuropeEemshaven, Netherlands53°25′31″N6°51′44″E / 53.4252171°N 6.8622574°E /53.4252171; 6.8622574EemshavenNetherlands (europe-west4)2014 - announced

2016 - launched

2018, 2019 - expansion

200 employees
EuropeFrankfurt, Germany50°07′21″N8°58′27″E / 50.1226299°N 8.9741678°E /50.1226299; 8.9741678[13]-Frankfurt (europe-west3)2022 - expanded[14]
EuropeFredericia, Denmark55°33′29.5″N9°39′20.8″E / 55.558194°N 9.655778°E /55.558194; 9.655778Fredericia-2018 - announced[15]

2020 - launched

€600M building costs
EuropeGhlin, Hainaut, Belgium50°28′09.6″N3°51′55.7″E / 50.469333°N 3.865472°E /50.469333; 3.865472Saint-GhislainBelgium (europe-west1)2007 - announced

2010 - launched

12 employees
EuropeHamina, Finland60°32′11.68″N27°7′1.21″E / 60.5365778°N 27.1170028°E /60.5365778; 27.1170028HaminaFinland

(europe-north1)

2009 - announced

2011 - first phase completed

2022 - expansion

6 buildings, 400 employees[16]
North AmericaHenderson (NV), USA36°03′20″N115°00′37″W / 36.0556248°N 115.0102258°W /36.0556248; -115.0102258HendersonLas Vegas (us-west4)2019 - announced[17]

2020 - launched

64-acres; $1.2B building costs[18][19]
AsiaHong Kong, Hong Kong-Hong Kong (asia-east2)2017 - announced[20]

2018 - launched[21]

AsiaInzai, Japan35°49′04″N140°07′57″E / 35.8177804°N 140.1323916°E /35.8177804; 140.1323916Inzai-2023 - launched
AsiaJakarta, Indonesia-Jakarta (asia-southeast2)2020 - launched[22]
AsiaKoto-Ku, Tokyo, Japan-Tokyo

(asia-northeast1)

2016 - launched[23]
North AmericaLeesburg (VA), USA39°3′6.38″N77°32′20.38″W / 39.0517722°N 77.5389944°W /39.0517722; -77.5389944Loudoun CountyN. Virginia (us-east4)2017 - announced[4][5]
North AmericaLenoir (NC), USA35°53′54.78″N81°32′50.58″W / 35.8985500°N 81.5473833°W /35.8985500; -81.5473833Lenoir-2007 - announced

2009 - launched

over 110 employees
AsiaLok Yang Way,Pioneer, Singapore1°19′26″N103°41′36″E / 1.3239859°N 103.693253°E /1.3239859; 103.693253[24]SingaporeSingapore (asia-southeast1)2022 - launched
EuropeLondon, UK-London

(europe-west2)

2017 - launched[25]
North AmericaLos Angeles (CA), USA-Los Angeles (us-west2)
EuropeMadrid, Spain40°31′10″N3°20′27″W / 40.5195333°N 3.3409366°W /40.5195333; -3.3409366-Madrid (europe-southwest1)2022 - launched[26]
PacificMelbourne, Australia-Melbourne

(australia-southeast2)

2021 - launched[27]
EuropeMiddenmeer, Noord-Holland, The Netherlands52°47′24″N5°01′45″E / 52.7901053°N 5.0292193°E /52.7901053; 5.0292193[28]MiddenmeerNetherlands (europe-west4)2019 - announced[29]
North AmericaMidlothian (TX), USA32°26′35″N97°03′44″W / 32.443170°N 97.062324°W /32.443170; -97.062324MidlothianDallas (us-south1)2019 - announced

2022 - launched[30]

375-acres; $600M building costs[31]
EuropeMilan, Italy-Milan (europe-west8)2022 - launched[32]
North AmericaMoncks Corner (SC), USA33°03′50.8″N80°02′36.1″W / 33.064111°N 80.043361°W /33.064111; -80.043361Berkeley CountySouth Carolina (us-east1)2007 - launched

2013 - expanded

150 employees
North AmericaMontreal, Quebec, Canada[33]-Montréal (northamerica-northeast1)2018 - launched[34]62.4-hectares; $600M building costs[35]
AsiaMumbai, India-Mumbai (asia-south1)2017 - launched[36]
North AmericaNew Albany (OH), USA40°03′41″N82°45′31″W / 40.0613352°N 82.7586278°W /40.0613352; -82.7586278New Albany-2019 - announced400-acres; $600M building costs[37][38]
AsiaOsaka, Japan-Osaka

(asia-northeast2)

2019 - launched[39]
South AmericaOsasco, São Paulo, Brazil-São Paulo (southamerica-east1)2017 - launched[40]
North AmericaPapillion (NE), USA41°08′00″N96°08′39″W / 41.1332915°N 96.144178°W /41.1332915; -96.144178Papillion-2019 - announced275-acres; $600M building costs[41][42]
EuropeParis, France-Paris (europe-west9)2022 - launched[43]
North AmericaPryor Creek (OK), USA36°14′28.1″N95°19′48.22″W / 36.241139°N 95.3300611°W /36.241139; -95.3300611Mayes County-2007 - announced

2012 - expanded

over 400 employees,[44] land atMidAmerica Industrial Park
South AmericaQuilicura, Santiago, Chile33°21′30.5″S70°41′50.4″W / 33.358472°S 70.697333°W /-33.358472; -70.697333Quilicura-2012 - announced

2015 - launched

up to 20 employees expected. A million dollar investment plan to increase capacity at Quilicura was announced in 2018.[45]
North AmericaReno (NV), USA39°30′04″N119°25′46″W / 39.5011558°N 119.4295537°W /39.5011558; -119.4295537Storey County-2017 - 1,210 acres of land bought in theTahoe Reno Industrial Center[46]

2018 - announced

2018 November - project approved by the state of Nevada[47][48]

North AmericaSalt Lake City (UT), USA-Salt Lake City (us-west3)2020 - launched[49]
AsiaSeoul, South Korea-Seoul

(asia-northeast3)

2020 - launched[50]
PacificSydney, Australia-Sydney

(australia-southeast1)

2017 - launched[51]
Middle EastTel Aviv, Israel[52]-Tel Aviv (me-west1)2022 - launched[53]
North AmericaThe Dalles (OR), USA45°37′57.04″N121°12′8.16″W / 45.6325111°N 121.2022667°W /45.6325111; -121.2022667The DallesOregon (us-west1)2006 - launched80 full-time employees
North AmericaToronto, Canada-Toronto (northamerica-northeast2)2021 - launched[54]
EuropeTurin, Italy45°08′48″N7°44′32″E / 45.146729°N 7.742147°E /45.146729; 7.742147-Turin (europe-west12)2023 - launched[55]
South AmericaVinhedo, São Paulo, BrazilSão Paulo (southamerica-east1)
EuropeWarsaw, Poland-Warsaw (europe-central2)2019 - announced

2021 - launched[56]

AsiaWenya, Jurong West, Singapore1°21′04.8″N103°42′35.2″E / 1.351333°N 103.709778°E /1.351333; 103.709778SingaporeSingapore (asia-southeast1)2011 - announced

2013 - launched

2015 - expanded

North AmericaWidows Creek (Bridgeport) (AL), USA34°54′48.4″N85°44′53.1″W / 34.913444°N 85.748083°W /34.913444; -85.748083[57]Jackson County-2018 - broke ground
EuropeZürich, Switzerland47°26′45″N8°12′39″E / 47.4459257°N 8.2109085°E /47.4459257; 8.2109085[58]-Zurich (europe-west6)2018 - announced

2019 - launched[59]

EuropeAustria2022 - announced[60]
EuropeBerlin, Germany[61]Berlin (europe-west10)2021 - announced[14] 2023 August - launched[62]
Middle EastDammam, Saudi Arabia2021 - announced[63]
EuropeAthens, Greece2022 - announced[60]
North AmericaKansas City, Missouri2019 - announced[64]
Middle EastKuwait2023 - announced[65]
AsiaMalaysia2022 - announced[66]
PacificAuckland, New Zealand2022 - announced[66]
EuropeOslo, Norway2022 - announced[60]
North AmericaQuerétaro, Mexico2022 - announced[67]
AfricaJohannesburg, South AfricaJohannesburg (africa-south1)2022 - announced[60] 2024 - launched
EuropeSweden2022 - announced[60]
AsiaTainan City, Taiwan-Taiwan

(asia-east1)

2019 September - announced[68][69][70]
AsiaThailand2022 - announced[66]
AsiaYunlin County, Taiwan-Taiwan (asia-east1)2020 September - announced[71]
North AmericaMesa (AZ), USAPhoenix (us-west8)2023 - construction started[72]
EuropeWaltham Cross, Hertfordshire, UK51°41′44″N0°02′55″W / 51.695666°N 0.048728°W /51.695666; -0.0487282024 January - announced[73]
South AmericaCanelones, Uruguay34°48′56″S55°59′44″W / 34.815481°S 55.995683°W /-34.815481; -55.9956832024 - construction started[74]

2026 - inauguration expected[75]

AsiaVisakhapatnam,Andhra Pradesh,IndiaAnnounced July 2025.[76]Google plans to invest US$15 billion to build a 1 GW data center campus in Visakhapatnam, including US$2 billion for renewable energy infrastructure. It is expected to be the company’s largest data center project in Asia.[77]

Hardware

[edit]

Original hardware

[edit]
Google's first production server rack, circa 1998

The original hardware (circa 1998) that was used by Google when it was located atStanford University included:[78]

  • Sun Microsystems Ultra II with dual 200 MHz processors, and 256 MB ofRAM. This was the main machine for the originalBackrub system.
  • 2 × 300 MHz dualPentium II servers donated byIntel, they included 512 MB of RAM and 10 × 9 GB hard drives between the two. It was on these that the main search ran.
  • F50 IBMRS/6000 donated byIBM, included 4 processors, 512 MB of memory and 8 × 9 GB hard disk drives.
  • Two additional boxes included 3 × 9 GB hard drives and 6 x 4 GB hard disk drives respectively (the original storage for Backrub). These were attached to the Sun Ultra II.
  • SSD disk expansion box with another 8 × 9 GB hard disk drives donated by IBM.
  • Homemade disk box which contained 10 × 9 GBSCSI hard disk drives.

Google Cluster

[edit]

The state of Google infrastructure in 2003 was described in a report byLuiz André Barroso,Jeff Dean, andUrs Hölzle as a "reliable computing infrastructure from clusters of unreliable commodity PCs".[79]

At the time, on average, a single search query read ~100MB of data, and consumed1010{\displaystyle \sim 10^{10}} CPU cycles. During peak time, Google served ~1000 queries per second. To handle this peak load, they built a compute cluster with ~15,000 commodity-class PCs instead of expensive supercomputer hardware to save money. To make up for the lower hardware reliability, they wrotefault tolerant software.

The structure of the cluster consists of five parts. Central Google Web servers (GWS) face the public Internet. Upon receiving a user request, the Google Web server communicates with a spell checker, an advertisement server, many index servers, many document servers. Each of the four parts responds to a part of the request, and the GWS assembles their responses and serves the final response to the user.

The raw documents were ~100 TB, and the index files were ~10 TB. The index files are sharded, and each shard is served by a "pool" of index servers. Similarly, the raw documents are also sharded. Each query to the index file results in a list of document IDs, which are then sent to the document servers to retrieve the title and the keyword-in-context snippets.

There were several CPU generations in use, ranging from single-processor 533 MHzIntel-Celeron-based servers to dual 1.4 GHz IntelPentium III. Each server contained one or more hard drives, 80GB each. Index servers have less disk space than document servers. Each rack had twoEthernet switches, one per side. The servers on each side interconnected via a 100-Mbps. Each switch had a ~250 MB/sec uplink to a central switch that connected to all racks.

The design objectives include:

  • Use low-reliability consumer hardware and make up for it with fault-tolerant software.
  • Maximize parallelism, such as by splitting a single document match lookup in a large index into aMapReduce over many small indices.
  • Partition index data and computation to minimize communication and evenly balance the load across servers, because the cluster is a large shared-memory machine.
  • Minimize system management overheads by developing all software in-house.
  • Pick hardware that maximizes performance/price, not absolute performance.
  • Pick hardware that has highthoroughput over highlatency. This is because queries are served with massive parallelism, with very few dependent steps and minimal communication between servers, so high latency does not matter.

Due to the massive parallelism, scaling up hardware scales up the thoroughput linearly, i.e. doubling the compute cluster doubles the number of queries servable per second.

The cluster is made of server racks at 2 configurations: 40 x1u per side with 2 sides, or 20 x2u per side with 2 sides. The power consumption is 10 kW per rack, at a density of 400 W/ft^2, consuming 10MWh per month, costing $1,500 per month.

Production hardware

[edit]

As of 2014, Google has used a heavily customized version ofDebianLinux. They migrated from a Red Hat-based system incrementally in 2013.[80]

The customization goal is to purchase CPU generations that offer the best performance per dollar, not absolute performance. How this is measured is unclear, but it is likely to incorporate running costs of the entire server, and CPU power consumption could be a significant factor.[81] Servers as of 2009–2010 consisted of custom-made open-top systems containing two processors (each with several cores[82]), a considerable amount of RAM spread over 8 DIMM slots housing double-height DIMMs, and at least two SATA hard disk drives connected through a non-standard ATX-sized power supply unit.[83] The servers were open top so more servers could fit into a rack. According to CNET and a book byJohn Hennessy, each server had a novel 12-volt battery to reduce costs and improve power efficiency.[82][84]

[needs update]According to Google, their global data center operation electrical power ranged between 500 and 681megawatts by 2010.[85][86] The combined processing power of these servers might have reached from 20 to 100petaflops in 2008.[87]

Network topology

[edit]

Details of the Google worldwide private networks are not publicly available, but Google publications[88][89] make references to the "Atlas Top 10" report that ranks Google as the third largest ISP behindLevel 3.

In order to run such a large network, with direct connections to as many ISPs as possible at the lowest possible cost, Google has a very openpeering policy.[90]

From this site, we can see that the Google network can be accessed from 67 public exchange points and 69 different locations across the world. As of May 2012, Google had 882 Gbit/s of public connectivity (not counting private peering agreements that Google has with the largest ISPs). This public network is used to distribute content to Google users as well as to crawl the internet to build its search indexes.The private side of the network is a secret, but a recent disclosure from Google[91] indicate that they use custom built high-radixswitch-routers (with a capacity of 128 × 10Gigabit Ethernet port) for thewide area network. Running no less than two routers per datacenter (for redundancy) we can conclude that the Google network scales in the terabit per second range (with two fully loaded routers the bi-sectional bandwidth amount to 1,280 Gbit/s).

These custom switch-routers are connected toDWDM devices tointerconnect data centers andpoint of presences (PoP) viadark fiber.

From a datacenter view, the network starts at the rack level, where19-inch racks are custom-made and contain 40 to 80 servers (20 to 40 1U servers on either side, while new servers are 2U rackmount systems.[92] Each rack has anEthernet switch). Servers are connected via a 1 Gbit/sEthernet link to the top of rack switch (TOR). TOR switches are then connected to agigabit cluster switch using multiple gigabit or ten gigabit uplinks.[93] The cluster switches themselves are interconnected and form the datacenter interconnect fabric (most likely using a dragonfly design rather than a classic butterfly or flattened butterfly layout[94]).

From an operation standpoint, when a client computer attempts to connect to Google, severalDNS servers resolvewww.google.com into multiple IP addresses viaRound Robin policy. Furthermore, this acts as the first level ofload balancing and directs the client to different Google clusters. A Google cluster has thousands ofservers, and once the client has connected to the server additional load balancing is done to send the queries to the least loaded web server. This makes Google one of the largest and most complexcontent delivery networks.[95]

Google has numerous data centers scattered around the world. At least 12 significant Google data center installations are located in the United States. The largest known centers are located inThe Dalles, Oregon;Atlanta, Georgia;Reston, Virginia;Lenoir, North Carolina; andMoncks Corner, South Carolina.[96] In Europe, the largest known centers are inEemshaven andGroningen in theNetherlands andMons,Belgium.[96] Google'sOceania Data Center is located inSydney,Australia.[97]

Data center network topology

[edit]

To supportfault tolerance, increase the scale of data centers and accommodate low-radix switches, Google has adopted various modifiedClos topologies in the past.[98]

Project 02

[edit]
Google data center in The Dalles, Oregon

One of the largest Google data centers is located in the town ofThe Dalles, Oregon, on theColumbia River, approximately 80 miles (129 km) fromPortland. Codenamed "Project 02", the complex was built in 2006 and is approximately the size of twoAmerican football fields, withcooling towers four stories high.[99][100] The site was chosen to take advantage of inexpensivehydroelectric power, and to tap into the region's largesurplus offiber optic cable, a remnant of thedot-com boom. A blueprint of the site appeared in 2008.[101]

Summa papermill

[edit]

In February 2009,Stora Enso announced that they had sold the Summa paper mill inHamina,Finland to Google for 40 million Euros.[102][103] Google invested 200 million euros on the site to build a data center and announced additional 150 million euro investment in 2012.[104][105] Google chose this location due to the availability and proximity of renewable energy sources.[106]

Floating data centers

[edit]
See also:Google barges

In 2013, the press revealed the existence of Google's floating data centers along the coasts of the states of California (Treasure Island's Building 3) and Maine. The development project was maintained under tight secrecy. The data centers are 250 feet long, 72 feet wide, 16 feet deep. The patent for an in-ocean data center cooling technology was bought by Google in 2009[107][108] (along with a wave-powered ship-based data center patent in 2008[109][110]). Shortly thereafter, Google declared that the two massive and secretly built infrastructures were merely "interactive learning centers, [...] a space where people can learn about new technology."[111]

Google halted work on the barges in late 2013 and began selling off the barges in 2014.[112][113]

Software

[edit]

Most of thesoftware stack that Google uses on their servers was developed in-house.[114] According to a well-known former Google employee in 2006,C++,Java,Python and (more recently)Go are favored over other programming languages.[115] For example, the back end of Gmail is written in Java and the back end of Google Search is written in C++.[116] Google has acknowledged that Python has played an important role from the beginning, and that it continues to do so as the system grows and evolves.[117]

The software that runs the Google infrastructure includes:[118]

Google has developed several abstractions which it uses for storing most of its data:[126]

  • Protocol Buffers – "Google's lingua franca for data",[127] a binary serialization format which is widely used within the company.
  • SSTable (Sorted Strings Table) – a persistent, ordered, immutable map from keys to values, where both keys and values are arbitrary byte strings. It is also used as one of the building blocks of Bigtable.[128]
  • RecordIO – a sequence of variable sized records.[126][129][130]

Software development practices

[edit]

Most operations are read-only. When an update is required, queries are redirected to other servers, so as to simplify consistency issues. Queries are divided into sub-queries, where those sub-queries may be sent to different ducts inparallel, thus reducing the latency time.[92]

To lessen the effects of unavoidablehardware failure, software is designed to befault tolerant. Thus, when a system goes down, data is still available on other servers, which increases reliability.

Search infrastructure

[edit]
Google data center inThe Dalles, Oregon

Index

[edit]

Like most search engines, Google indexes documents by building a data structure known asinverted index. Such an index obtains a list of documents by a query word. The index is very large due to the number of documents stored in the servers.[95]

The index is partitioned by document IDs into many pieces calledshards. Each shard isreplicated onto multiple servers. Initially, the index was being served fromhard disk drives, as is done in traditionalinformation retrieval (IR) systems. Google dealt with the increasing query volume by increasing number of replicas of each shard and thus increasing number of servers. Soon they found that they had enough servers to keep a copy of the whole index in main memory (although with low replication or no replication at all), and in early 2001 Google switched to anin-memory index system. This switch "radically changed many design parameters" of their search system, and allowed for a significant increase in throughput and a large decrease in latency of queries.[131]

In June 2010, Google rolled out a next-generation indexing and serving system called "Caffeine" which can continuously crawl and update the search index. Previously, Google updated its search index in batches using a series ofMapReduce jobs. The index was separated into several layers, some of which were updated faster than the others, and the main layer wouldn't be updated for as long as two weeks. With Caffeine, the entire index is updated incrementally on a continuous basis. Later Google revealed a distributed data processing system called "Percolator"[132] which is said to be the basis of Caffeine indexing system.[124][133]

Server types

[edit]

Google's server infrastructure is divided into several types, each assigned to a different purpose:[92][95][134][135][136]

  • Web servers coordinate the execution of queries sent by users, then format the result into anHTML page. The execution consists of sending queries to index servers, merging the results, computing their rank, retrieving a summary for each hit (using the document server), asking for suggestions from the spelling servers, and finally getting a list of advertisements from the ad server.
  • Data-gathering servers are permanently dedicated tospidering the Web. Google's web crawler is known as GoogleBot. They update the index and document databases and apply Google's algorithms to assign ranks to pages.
  • Each index server contains a set of index shards. They return a list of document IDs ("docid"), such that documents corresponding to a certain docid contain the query word. These servers need less disk space, but suffer the greatest CPU workload.
  • Document servers store documents. Each document is stored on dozens of document servers. When performing a search, a document server returns a summary for the document based on query words. They can also fetch the complete document when asked. These servers need more disk space.
  • Ad servers manage advertisements offered by services likeAdWords andAdSense.
  • Spelling servers make suggestions about the spelling of queries.

There are also "canary requests", whereby a request is first sent to one or two leaf servers to see if the response time is reasonable. If not, then the request fails. This provides security.[137]

Security

[edit]
External videos
YouTube logo
YouTube
video iconGoogle Data Center Security: 6 Layers Deep

In October 2013,The Washington Post reported that the U.S.National Security Agency intercepted communications between Google's data centers, as part of a program namedMUSCULAR.[138][139] This wiretapping was made possible because, at the time, Google did not encrypt data passed inside its own network.[140] This was rectified when Google began encrypting data sent between data centers in 2013.[141]

Environmental impact

[edit]
Google data center inMayes County, Oklahoma atMidAmerica Industrial Park

Some of Google's most efficient data centers in 2012 ran at 35 °C (95 °F) using only fresh air cooling, requiring no electrically powered air conditioning.[142][needs update]

In December 2016, Google announced that—starting in 2017—it would purchase enough renewable energy to match 100% of the energy usage of its data centers and offices. The commitment will make Google "the world's largest corporate buyer of renewable power, with commitments reaching 2.6 gigawatts (2,600 megawatts) of wind and solar energy".[143][144][145]

In 2025, Google agreed to pay to restart the 600 MWDuane Arnold nuclear power station in Iowa by 2029.[146]

References

[edit]
  1. ^"How Many Servers Does Google Have?".Data Center Knowledge. March 16, 2017. RetrievedSeptember 20, 2018.
  2. ^"Google data centers, locations". RetrievedJuly 21, 2014.
  3. ^"ISO/IEC 27001 - Compliance".Google Cloud. RetrievedJuly 11, 2023.
  4. ^abReport, Times-Mirror Staff (June 19, 2019)."Google 'caps off' $600M investment in Loudoun County".LoudounTimes.com. RetrievedFebruary 12, 2021.
  5. ^ab"Google Plans 2 Loudoun Data Centers".LoudounNow.com. November 29, 2017. RetrievedApril 12, 2023.
  6. ^Arellano (CIPER), Alberto (May 25, 2020)."Las zonas oscuras de la evaluación ambiental que autorizó "a ciegas" el megaproyecto de Google en Cerrillos".CIPER Chile (in Spanish). RetrievedJuly 10, 2023.
  7. ^"Google instalará un nuevo data center en Chile".diarioeldia.cl (in Spanish). Archived fromthe original on September 22, 2021. RetrievedSeptember 28, 2019.
  8. ^"Una nueva Región de Google Cloud en Santiago, todo el potencial de la nube ahora más cerca".Google. December 1, 2021. RetrievedJuly 10, 2023.
  9. ^"A Google Cloud region now available in Columbus, Ohio".Google Cloud Blog. RetrievedJuly 10, 2023.
  10. ^"Namaste, India. Our new cloud region in Delhi NCR is now live".cloudonair.withgoogle.com.
  11. ^"New Google Cloud region now open in Qatar".Google Cloud Blog. RetrievedJuly 10, 2023.
  12. ^"Dublin, Ireland – Data Centers – Google".www.google.com. RetrievedApril 2, 2019.
  13. ^Barazza (September 2, 2021)."PROJEKT IN HANAU: RECHENZENTRUM FÜR GOOGLE ENTWICKELT".NDC-GARBE.
  14. ^ab"Google Cloud investing in Germany with new infrastructure and sustainable energy".Google Cloud Blog. RetrievedJuly 10, 2023.
  15. ^"Breaking ground for Google's first data center in Denmark".Google. November 20, 2018. RetrievedJuly 10, 2023.
  16. ^"Google gets green light for its sixth data center in Hamina, Finland". September 11, 2020.
  17. ^"Investing in Google infrastructure, investing in Nevada".Google Cloud Blog. RetrievedJuly 10, 2023.
  18. ^Torres-Cortez, Ricardo (September 16, 2020)."Google to invest additional $600M at Henderson data center – Las Vegas Sun Newspaper".lasvegassun.com. RetrievedFebruary 12, 2021."With this latest announcement, Google will bring their total investment in the city of Henderson to $1.2 billion," said Mayor Debra March in the release
  19. ^Baxtel."Google Henderson NV Data Center".baxtel.com. RetrievedFebruary 12, 2021.
  20. ^"Coming soon: GCP's Hong Kong region".Google Cloud Blog. RetrievedJuly 10, 2023.
  21. ^"Growing our presence in Asia Pacific: New GCP regions in Hong Kong and Jakarta".Google Cloud Blog. RetrievedJuly 10, 2023.
  22. ^"The new Google Cloud region in Jakarta is now open".Google Cloud Blog. RetrievedSeptember 15, 2021.
  23. ^"Google Cloud Platform Tokyo region now open for business".Google Cloud Blog. RetrievedJuly 10, 2023.
  24. ^"Google - Singapore".www.streetdirectory.com. RetrievedJuly 11, 2023.
  25. ^"Google Cloud Platform now open in London".Google Cloud Blog. RetrievedJuly 10, 2023.
  26. ^"New Google Cloud region in Madrid, Spain now open".Google Cloud Blog. RetrievedJuly 10, 2023.
  27. ^"The Google Cloud region in Melbourne is now open".Google Cloud Blog. RetrievedJuly 10, 2023.
  28. ^"Welcome".Agriport A7 (in Dutch). RetrievedJuly 10, 2023.
  29. ^"Google to Spend $1.1 Billion on New Data Centers in Netherlands".Data Center Knowledge. June 24, 2019.
  30. ^"A Google Cloud region now available in Dallas, Texas".Google Cloud Blog. RetrievedJuly 10, 2023.
  31. ^"Google's massive $600M data center takes shape in Ellis County as tech giant ups Texas presence".Dallas News. June 14, 2019. RetrievedFebruary 12, 2021.
  32. ^"New Google Cloud region in Milan, Italy now open".Google Cloud Blog. RetrievedJuly 10, 2023.
  33. ^"Project to expand Google's activities in Quebec - Future computer data center in Beauharnois".www.quebec.ca (in French). RetrievedMay 10, 2021.
  34. ^"GCP arrives in Canada with launch of Montréal region".Google Cloud Blog. RetrievedJuly 10, 2023.
  35. ^"Google's ongoing commitment in Quebec".Google Blog - Company News. RetrievedFebruary 12, 2021.
  36. ^Stiver, Dave (November 1, 2017)."GCP arrives in India with launch of Mumbai region".Google Cloud Blog. RetrievedJuly 30, 2019.
  37. ^Williams, Mark."Google joins New Albany high-tech crowd with $600 million data center".The Columbus Dispatch. Archived fromthe original on December 30, 2020. RetrievedFebruary 12, 2021.
  38. ^"New Albany, Ohio – Data Centers – Google".Google Data Centers. RetrievedFebruary 12, 2021.
  39. ^"Google Cloud launches new Osaka region to support growing customer base in Japan".Google Cloud Blog. RetrievedJuly 10, 2023.
  40. ^"GCP arrives in South America with launch of São Paulo region!".Google Cloud Blog. RetrievedJuly 10, 2023.
  41. ^"Google confirms it is behind $600m Papillion data center project".www.datacenterdynamics.com. October 4, 2019. RetrievedFebruary 12, 2021.
  42. ^"Papillion, Nebraska – Data Centers – Google".Google Data Centers. RetrievedFebruary 12, 2021.
  43. ^"Google Cloud region in Paris France now open".Google Cloud Blog. RetrievedJuly 10, 2023.
  44. ^Dawn-Hiscox, Tanwen (February 20, 2018)."Google to spend m on Pryor data center expansion".Data Centre Dynamics.Archived from the original on April 23, 2019. RetrievedApril 23, 2019.
  45. ^"Google ha decido de invertir millones de dólares en su centro de datos en Chile".Newtechmag.net (in Spanish). September 28, 2018. Archived fromthe original on March 30, 2019. RetrievedDecember 8, 2018.
  46. ^Tanwen Dawn-Hiscox (April 18, 2017)."Google is planning a massive data center in Nevada".Datacenterdynamics.com. RetrievedDecember 8, 2018.
  47. ^Jason Hidalgo (November 16, 2018)."Nevada approves Google's M data center near Las Vegas, M in tax incentives".Rgj.com. RetrievedDecember 8, 2018.
  48. ^Jason Hidalgo (September 16, 2020)."Google to invest $600 million in data center near Reno, gets tax break".Reno Gazette Journal. RetrievedOctober 26, 2020.With our new data center in Storey County and our expanded investment in our Henderson site, Google will have two facilities in Nevada, bringing our total investment to over $1.88 billion.
  49. ^"Google Cloud region in Salt Lake City now open".Google Cloud Blog. RetrievedJuly 10, 2023.
  50. ^"New GCP Region in Seoul".Google Cloud Blog. RetrievedJuly 10, 2023.
  51. ^"Google Cloud expands to Australia with new Sydney region".Google Cloud Blog. RetrievedJuly 10, 2023.
  52. ^"Google launches GCP region in Tel Aviv, Israel".datacenterdynamics.com. October 19, 2022.
  53. ^"Google Cloud region in Tel Aviv Israel now open".Google Cloud Blog. RetrievedJuly 10, 2023.
  54. ^"Google Cloud Toronto region now open".Google Cloud Blog. RetrievedJuly 10, 2023.
  55. ^"New Google Cloud region in Turin, Italy now open".Google Cloud Blog. RetrievedJuly 10, 2023.
  56. ^"Google to Build Cloud Data Centers in Poland".Data Center Knowledge. September 27, 2019.
  57. ^"Google kicks off construction on M Alabama data center".Made in Alabama. April 9, 2018. RetrievedAugust 19, 2019.
  58. ^"Die Schweizer Google Cloud Platform zieht zu Green in den Aargau".Swiss IT Magazine (in German). RetrievedJuly 10, 2023.
  59. ^<<Cite web|url=https://www.datacenterknowledge.com/google-alphabet/google-building-cloud-data-centers-close-swiss-banks%7Ctitle=Google[permanent dead link] Building Cloud Data Centers Close to Swiss Banks||
  60. ^abcde"Introducing 5 new Google Cloud regions".Google Cloud Blog. RetrievedJuly 10, 2023.
  61. ^"Google plant ein großes Rechenzentrum südlich des Flughafens BER".Der Spiegel (in German). September 30, 2022.ISSN 2195-1349. RetrievedJuly 10, 2023.
  62. ^"Berlin-Brandenburg Google Cloud region is now open".Google Cloud Blog. RetrievedFebruary 3, 2024.
  63. ^"Google Cloud Platform region updates".Google Cloud Blog. RetrievedJuly 10, 2023.
  64. ^"Google affiliate's latest move signals selection of KC for $600M data center".www.bizjournals.com. RetrievedFebruary 12, 2021.
  65. ^"Bringing a new Google Cloud region to Kuwait".Google Cloud Blog. RetrievedJuly 10, 2023.
  66. ^abc"Announcing new Google Cloud regions in Asia Pacific".Google Cloud Blog. RetrievedJuly 10, 2023.
  67. ^"Announcing a new Google Cloud region in Mexico".Google Cloud Blog. RetrievedJuly 10, 2023.
  68. ^"Google purchases land for new data center in Tainan".Taipei Times. September 12, 2019. RetrievedDecember 20, 2019.
  69. ^"Google to set up data center in Tainan".Focus Taiwan. September 11, 2019. RetrievedDecember 20, 2019.
  70. ^"Google to set up second data center in Taiwan".Taiwan News. September 11, 2019. RetrievedDecember 20, 2019.
  71. ^"Google confirms plans to build 3rd data center in Taiwan".Taiwan News. September 3, 2020. RetrievedSeptember 3, 2020.
  72. ^"Our work to build a more sustainable future in Arizona".Google. March 14, 2024. RetrievedOctober 5, 2024.
  73. ^"Our $1 billion investment in a new UK data centre".Google. January 18, 2024. RetrievedJanuary 19, 2024.
  74. ^"A new data center in Latin America".Google. August 29, 2024. RetrievedOctober 5, 2024.
  75. ^"Google y la piedra fundamental (comenzó la construcción del centro de datos en Parque de las Ciencias)" (in Spanish). August 30, 2024. RetrievedOctober 5, 2024.
  76. ^"Google to invest $15 billion in India for Asia's biggest AI datacentre project: Report".The Times of India. July 30, 2025.ISSN 0971-8257.
  77. ^Varadhan, Sudarshan; Varadhan, Sudarshan (July 30, 2025)."Google to invest $6 billion in southern India data centre, sources say".Reuters. RetrievedAugust 1, 2025.
  78. ^""Google Stanford Hardware"". Archived from the original on February 9, 1999. RetrievedMarch 23, 2017.{{cite web}}: CS1 maint: bot: original URL status unknown (link).Stanford University (provided byInternet Archive). Retrieved on July 10, 2006.
  79. ^Barroso, L.A.; Dean, J.; Holzle, U. (March 2003). "Web search for a planet: The Google cluster architecture".IEEE Micro.23 (2):22–28.Bibcode:2003IMicr..23b..22B.doi:10.1109/MM.2003.1196112.ISSN 1937-4143.
  80. ^Merlin, Marc (2013)."Case Study: Live upgrading many thousand of servers from an ancient Red Hat distribution to a 10 year newer Debian based one"(PDF).Linux Foundation. RetrievedJune 9, 2017.
  81. ^Tawfik Jelassi; Albrecht Enders (2004). "Case study 16 — Google".Strategies for E-business. Pearson Education. p. 424.ISBN 978-0-273-68840-2.
  82. ^abComputer Architecture, Fifth Edition: A Quantitative Approach,ISBN 978-0123838728; Chapter Six; 6.7 "A Google Warehouse-Scale Computer"page 471 "Designing motherboards that only need a single 12-volt supply so that the UPS function could be supplied by standard batteries associated with each server"
  83. ^Google's secret power supplies onYouTube
  84. ^Google uncloaks once-secret server, April 1, 2009.
  85. ^"Google Sustainability".Google Sustainability.
  86. ^"Analytics Press Growth in data center electricity use 2005 to 2010". Archived fromthe original on January 11, 2012. RetrievedMay 22, 2012.
  87. ^Google Surpasses Supercomputer Community, Unnoticed?Archived December 5, 2008, at theWayback Machine, May 20, 2008.
  88. ^Lam, Cedric F.; Liu, Hong; Koley, Bikash; Zhao, Xiaoxue; Kamalov, Valey; Gill, Vijay (2010), "Fiber Optic Communication Technologies: What's Needed for Datacenter Network Operations",Research, vol. 48
  89. ^Lam, Cedric F. (2010),FTTH look ahead — technologies & architectures(PDF), p. 4
  90. ^"kumara ASN15169",Peering DB
  91. ^"Urs Holzle",Speakers, Open Network Summit, archived fromthe original on May 10, 2012, retrievedMay 22, 2012
  92. ^abcWeb Search for a Planet: The Google Cluster Architecture (Luiz André Barroso, Jeffrey Dean, Urs Hölzle)
  93. ^"Warehouse size computers"(PDF). Archived fromthe original(PDF) on June 19, 2012. RetrievedMay 22, 2012.
  94. ^Denis Abt High Performance Datacenter Networks: Architectures, Algorithms, and Opportunities
  95. ^abcFiach Reid (2004)."Case Study: The Google search engine".Network Programming in .NET. Digital Press. pp. 251–253.ISBN 978-1-55558-315-6.
  96. ^abRich Miller (March 27, 2008)."Google Data Center FAQ".Data Center Knowledge. Archived fromthe original on March 13, 2009. RetrievedMarch 15, 2009.
  97. ^Brett Winterford (March 5, 2010)."Found: Google Australia's secret data network".ITNews. RetrievedMarch 20, 2010.
  98. ^Singh, Arjun; Ong, Joon; Agarwal, Amit; Anderson, Glen; Armistead, Ashby; Bannon, Roy; Boving, Seb; Desai, Gaurav; Felderman, Bob; Germano, Paulie; Kanagala, Anand (2015)."Jupiter Rising: A Decade of Clos Topologies and Centralized Control in Google's Datacenter Network".Proceedings of the 2015 ACM Conference on Special Interest Group on Data Communication. pp. 183–197.doi:10.1145/2785956.2787508.ISBN 978-1-4503-3542-3.S2CID 2817692.
  99. ^Markoff, John; Hansell, Saul. "Hiding in Plain Sight, Google Seeks More Power."New York Times. June 14, 2006. Retrieved on October 15, 2008.
  100. ^Google "The Dalles, Oregon Data Center" Retrieved on January 3, 2011.
  101. ^Strand, Ginger."Google Data Center"Harper's Magazine. March 2008. Retrieved on October 15, 2008.Archived August 30, 2012, at theWayback Machine
  102. ^"Stora Enso divests Summa Mill premises in Finland for million".Stora Enso. February 12, 2009. Archived fromthe original on April 13, 2009. RetrievedDecember 2, 2009.
  103. ^[dead link]"Stooora yllätys: Google ostaa Summan tehtaan".Kauppalehti (in Finnish). Helsinki. February 12, 2009. Archived fromthe original on February 14, 2009. RetrievedFebruary 12, 2009.
  104. ^"Google investoi 200 miljoonaa euroa Haminaan".Taloussanomat (in Finnish). Helsinki. February 4, 2009. RetrievedMarch 15, 2009.
  105. ^"Hamina, Finland". RetrievedApril 23, 2018.
  106. ^Finland – First Choice for Siting Your Cloud Computing Data Center.Archived July 6, 2013, at theWayback Machine Accessed August 4, 2010.
  107. ^Rory Carroll (October 30, 2013)."Google's worst-kept secret: floating data centers off US coasts".Theguardian.com. RetrievedDecember 8, 2018.
  108. ^Rich Miller (April 29, 2009)."Google Gets Patent for Data Center Barges".Datacenterknowledge.com. RetrievedDecember 8, 2018.
  109. ^Martin Lamonica (September 8, 2008)."Google files patent for wave-powered floating data center".Cnet.com. RetrievedDecember 8, 2018.
  110. ^"Google's ship based datacenter patent application surfaces".Datacenterdynamics.com. September 7, 2008. RetrievedDecember 8, 2018.
  111. ^"Google barge mystery solved: they're for 'interactive learning centers'".Theguardian.com. November 6, 2013. RetrievedDecember 8, 2018.
  112. ^Brandon Bailey (August 1, 2014)."Google confirms selling a mystery barge".San Jose Mercury News. RetrievedApril 7, 2015.
  113. ^Chris Morran (November 7, 2014)."What Happened To Those Google Barges?".Consumerist. RetrievedJanuary 15, 2017.
  114. ^Mark Levene (2005).An Introduction to Search Engines and Web Navigation. Pearson Education. p. 73.ISBN 978-0-321-30677-7.
  115. ^"Python Status Update". Artima. January 10, 2006. RetrievedFebruary 17, 2012.
  116. ^"Warning".Panela. Blog-city. Archived fromthe original on December 28, 2011. RetrievedFebruary 17, 2012.
  117. ^"Quotes about Python". Python. RetrievedFebruary 17, 2012.
  118. ^"Google Architecture". High Scalability. November 22, 2008. RetrievedFebruary 17, 2012.
  119. ^abcFikes, Andrew (July 29, 2010), "Storage Architecture and Challenges",TechTalk(PDF)[permanent dead link]
  120. ^"Colossus: Successor to the Google File System (GFS)". SysTutorials. November 29, 2012. RetrievedMay 10, 2016.
  121. ^Dean, Jeffrey 'Jeff' (2009), "Design, Lessons and Advice from Building Large Distributed Systems",Ladis(keynote talk presentation), Cornell
  122. ^Shute, Jeffrey 'Jeff'; Oancea, Mircea; Ellner, Stephan; Handy, Benjamin 'Ben'; Rollins, Eric; Samwel, Bart; Vingralek, Radek; Whipkey, Chad; Chen, Xin; Jegerlehner, Beat; Littlefield, Kyle; Tong, Phoenix (2012), "F1 — the Fault-Tolerant Distributed RDBMS Supporting Google's Ad Business",Research,Sigmod, archived fromthe original(presentation) on June 9, 2017, retrievedSeptember 18, 2012{{citation}}: CS1 maint: location missing publisher (link)
  123. ^"Google alums rev up a new search engine".Los Angeles Times. July 28, 2008. RetrievedSeptember 15, 2021.
  124. ^abThe Register.Google Caffeine jolts worldwide search machine
  125. ^"Google official release note". RetrievedSeptember 28, 2013.
  126. ^ab"Google Developing Caffeine Storage System | TechWeekEurope UK". Eweekeurope.co.uk. August 18, 2009. Archived fromthe original on November 15, 2011. RetrievedFebruary 17, 2012.
  127. ^"Developer Guide – Protocol Buffers – Google Code". RetrievedFebruary 17, 2012.
  128. ^Chang, Fay; Dean, Jeffrey; Ghemawat, Sanjay; Hsieh, Wilson C.; Wallach, Deborah A.; Burrows, Mike; Chandra, Tushar; Fikes, Andrew; Gruber, Robert E. (2006)."Bigtable: A Distributed Storage System for Structured Data"(PDF).Archived(PDF) from the original on June 4, 2025.
  129. ^windley on (June 24, 2008)."Phil Windley's Technometria | Velocity 08: Storage at Scale". Windley.com. RetrievedFebruary 17, 2012.
  130. ^"Message limit – Protocol Buffers | Google Groups". RetrievedFebruary 17, 2012.
  131. ^"Jeff Dean's keynote at WSDM 2009"(PDF). RetrievedFebruary 17, 2012.
  132. ^Daniel Peng, Frank Dabek. (2010).Large-scale Incremental Processing Using Distributed Transactions and Notifications. Proceedings of the 9th USENIX Symposium on Operating Systems Design and Implementation.
  133. ^The Register.Google Percolator – global search jolt sans MapReduce comedown
  134. ^Chandler Evans (2008). "Google Platform".Future of Google Earth. Madison Publishing Company. p. 299.ISBN 978-1-4196-8903-1.
  135. ^Chris Sherman (2005). "How Google Works".Google Power. McGraw-Hill Professional. pp. 10–11.ISBN 978-0-07-225787-8.
  136. ^Michael Miller (2007). "How Google Works".Googlepedia. Pearson Technology Group. pp. 17–18.ISBN 978-0-7897-3639-0.
  137. ^Dean, Jeffrey; Barroso, Luiz André (February 2013)."The tail at scale".Communications of the ACM.56 (2):74–80.doi:10.1145/2408776.2408794.ISSN 0001-0782.
  138. ^Gellman, Barton; Soltani, Ashkan (October 30, 2013)."NSA infiltrates links to Yahoo, Google data centers worldwide, Snowden documents say".The Washington Post. RetrievedNovember 1, 2013.
  139. ^Savage, Charlie; Miller, Claire Cain; Perlroth, Nicole (October 30, 2013)."N.S.A. Said to Tap Google and Yahoo Abroad".The New York Times. RetrievedMarch 9, 2017.
  140. ^Gallagher, Sean (October 31, 2013)."How the NSA's MUSCULAR tapped Google's and Yahoo's private networks".Ars Technica.Condé Nast. RetrievedMarch 9, 2017.
  141. ^Miller, Claire Cain (October 31, 2013)."Angry Over U.S. Surveillance, Tech Giants Bolster Defenses".The New York Times. RetrievedMarch 9, 2017.
  142. ^Humphries, Matthew (March 27, 2012)."Google's most efficient data center runs at 95 degrees".geek.com. Archived fromthe original on June 13, 2016. RetrievedJune 13, 2016.
  143. ^Hölzle, Urs (December 6, 2016)."We're set to reach 100% renewable energy — and it's just the beginning".The Keyword Google Blog. RetrievedDecember 8, 2016.
  144. ^Statt, Nick (December 6, 2016)."Google just notched a big victory in the fight against climate change".The Verge.Vox Media. RetrievedDecember 8, 2016.
  145. ^Etherington, Darrell (December 7, 2016)."Google says it will hit 100% renewable energy by 2017".TechCrunch.AOL. RetrievedDecember 8, 2016.
  146. ^"NextEra Energy partners with Google to restart Iowa nuclear plant". Reuters. October 27, 2025. RetrievedOctober 29, 2025.

Further reading

[edit]

External links

[edit]
a subsidiary ofAlphabet
Company
Divisions
Subsidiaries
Active
Defunct
Programs
Events
Infrastructure
People
Current
Former
Criticism
General
Incidents
Other
Software
A–C
D–N
O–Z
Operating systems
Machine learning models
Neural networks
Computer programs
Formats and codecs
Programming languages
Search algorithms
Domain names
Typefaces
A
B
C
D
E
F
G
H
I
J
K
L
M
N
O
P
Q
R
S
T
U
V
W
Y
Hardware
Pixel
Smartphones
Smartwatches
Tablets
Laptops
Other
Nexus
Smartphones
Tablets
Other
Other
Advertising
Antitrust
Intellectual
property
Privacy
Other
Related
Concepts
Products
Android
Street View coverage
YouTube
Other
Documentaries
Books
Popular culture
Other
Retrieved from "https://en.wikipedia.org/w/index.php?title=Google_data_centers&oldid=1321056481"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp