
Thehistory of supercomputing goes back to the 1960s when a series ofcomputers atControl Data Corporation (CDC) were designed bySeymour Cray to use innovative designs and parallelism to achieve superior computational peak performance.[1] TheCDC 6600, released in 1964, is generally considered the firstsupercomputer.[2][3] However, some earlier computers were considered supercomputers for their day such as the 1954IBM NORC and 1955AN/FSQ-7vacuum tube computers in the 1950s,[4][5] and in the early 1960s, theUNIVAC LARC (1960),[6] theIBM 7030 Stretch (1962),[7] and theManchesterAtlas (1962), all[specify] of which were of comparable power.[citation needed]
While the supercomputers of the 1980s used only a few processors, in the 1990s, machines with thousands of processors began to appear both in the United States and in Japan, setting new computational performance records.
By the end of the 20th century, massively parallel supercomputers with thousands of "off-the-shelf" processors similar to those found in personal computers were constructed and broke through theteraFLOPS computational barrier.
Progress in the first decade of the 21st century was dramatic and supercomputers with over 60,000 processors appeared, reaching petaFLOPS performance levels.
The term "Super Computing" was first used in theNew York World in 1929[8] to refer to large custom-builttabulators thatIBM had made forColumbia University.[9]
There were several lines of second generation computers that were substantially faster than most contemporary mainframes. These included
The second generation saw the introduction of features intended to supportmultiprogramming andmultiprocessor configurations, including master/slave (supervisor/problem) mode, storage protection keys, limit registers, protection associated with address translation, andatomic instructions.
In 1957, a group of engineers leftSperry Corporation to formControl Data Corporation (CDC) inMinneapolis, Minnesota.Seymour Cray left Sperry a year later to join his colleagues at CDC.[1] In 1960, Cray completed theCDC 1604, one of the first generation of commercially successfultransistorized computers and at the time of its release, the fastest computer in the world.[10] However, the sole fully transistorizedHarwell CADET was operational in 1951, and IBM delivered its commercially successful transistorizedIBM 7090 in 1959.

Around 1960, Cray decided to design a computer that would be the fastest in the world by a large margin. After four years of experimentation along with Jim Thornton, and Dean Roush and about 30 other engineers, Cray completed theCDC 6600 in 1964. Cray switched from germanium to silicon transistors, built byFairchild Semiconductor, that used the planar process. These did not have the drawbacks of the mesa silicon transistors. He ran them very fast, and thespeed of light restriction forced a very compact design with severe overheating problems, which were solved by introducing refrigeration, designed by Dean Roush.[11] The 6600 outperformed the industry's prior recordholder, theIBM 7030 Stretch,[clarification needed] by a factor of three.[12][13] With performance of up to three megaFLOPS,[14][15] it was dubbed asupercomputer and defined the supercomputing market when two hundred computers were sold at $9 million each.[10][16]
The 6600 gained speed by "farming out" work to peripheral computing elements, freeing the CPU (Central Processing Unit) to process actual data. The MinnesotaFORTRAN compiler for the machine was developed by Liddiard and Mundstock at theUniversity of Minnesota and with it the 6600 could sustain 500 kiloflops on standard mathematical operations.[17] In 1968, Cray completed theCDC 7600, again the fastest computer in the world.[10] At 36 MHz, the 7600 had 3.6 times theclock speed of the 6600, but ran significantly faster due to other technical innovations. They sold only about 50 of the 7600s, not quite a failure. Cray left CDC in 1972 to form his own company.[10] Two years after his departure CDC delivered theSTAR-100, which at 100 megaflops was three times the speed of the 7600. Along with theTexas Instruments ASC, the STAR-100 was one of the first machines to usevector processing—the idea having been inspired around 1964 by theAPL programming language.[18][19]
In 1956, a team atManchester University in the United Kingdom began development ofMUSE—a name derived frommicrosecondengine—with the aim of eventually building a computer that could operate at processing speeds approaching one microsecond per instruction, about one millioninstructions per second.[20]Mu (the name of the Greek letterμ) is a prefix in the SI and other systems of units denoting a factor of 10−6 (one millionth).
At the end of 1958,Ferranti agreed to collaborate with Manchester University on the project, and the computer was shortly afterwards renamedAtlas, with the joint venture under the control ofTom Kilburn. The first Atlas was officially commissioned on 7 December1962—nearly three years before the Cray CDC 6600 supercomputer wasintroduced—as one of the world's firstsupercomputers. It was considered at the time of its commissioning to be the most powerful computer in the world, equivalent to fourIBM 7094s. It was said that whenever Atlas went offline half of the United Kingdom's computer capacity was lost.[21] The Atlas pioneeredvirtual memory andpaging as a way to extend its working memory by combining its 16,384 words of primarycore memory with an additional 96K words of secondarydrum memory.[22] Atlas also pioneered theAtlas Supervisor, "considered by many to be the first recognizable modernoperating system".[21]

Four years after leaving CDC, Cray delivered the 80 MHzCray-1 in 1976, and it became the most successful supercomputer in history.[19][23] The Cray-1, which used integrated circuits with two gates per chip, was avector processor. It introduced a number of innovations, such aschaining, in which scalar and vector registers generate interim results that can be used immediately, without additional memory references which would otherwise reduce computational speed.[11][24] TheCray X-MP (designed bySteve Chen) was released in 1982 as a 105 MHz shared-memoryparallelvector processor with better chaining support and multiple memory pipelines. All three floating-point pipelines on the X-MP could operate simultaneously.[24] By 1983 Cray and Control Data were supercomputer leaders; despite its lead in the overall computer market, IBM was unable to produce a profitable competitor.[25]
TheCray-2, released in 1985, was a four-processorliquid cooled computer totally immersed in a tank ofFluorinert, which bubbled as it operated.[11] It reached 1.9 gigaflops and was the world's fastest supercomputer, and the first to break the gigaflop barrier.[26] The Cray-2 was a totally new design. It did not use chaining and had a high memory latency, but used much pipelining and was ideal for problems that required large amounts of memory.[24] The software costs in developing a supercomputer should not be underestimated, as evidenced by the fact that in the 1980s the cost for software development at Cray came to equal what was spent on hardware.[27] That trend was partly responsible for a move away from the in-house,Cray Operating System toUNICOS based onUnix.[27]
TheCray Y-MP, also designed by Steve Chen, was released in 1988 as an improvement of the X-MP and could have eightvector processors at 167 MHz with a peak performance of 333 megaflops per processor.[24] In the late 1980s, Cray's experiment on the use ofgallium arsenide semiconductors in theCray-3 did not succeed. Seymour Cray began to work on amassively parallel computer in the early 1990s, but died in a car accident in 1996 before it could be completed. Cray Research did, however, produce such computers.[23][11]
TheCray-2 which set the frontiers of supercomputing in the mid to late 1980s had only 8 processors. In the 1990s, supercomputers with thousands of processors began to appear. Another development at the end of the 1980s was the arrival of Japanese supercomputers, some of which were modeled after the Cray-1.
During the first half of theStrategic Computing Initiative, some massively parallel architectures were proven to work, such as theWARP systolic array, message-passingMIMD like theCosmic Cube hypercube,SIMD like theConnection Machine, etc. In 1987, a TeraOPS Computing Technology Program was proposed, with a goal of achieving 1 teraOPS (a trillion operations per second) by 1992, which was considered achievable by scaling up any of the previously proven architectures.[28]

TheSX-3/44R was announced byNEC Corporation in 1989 and a year later earned the fastest-in-the-world title with a four-processor model.[29] However, Fujitsu'sNumerical Wind Tunnel supercomputer used 166 vector processors to gain the top spot in 1994. It had a peak speed of 1.7 gigaflops per processor.[30][31] TheHitachi SR2201 obtained a peak performance of 600 gigaflops in 1996 by using 2,048 processors connected via a fast three-dimensionalcrossbar network.[32][33][34]
In the same timeframe theIntel Paragon could have 1,000 to 4,000Intel i860 processors in various configurations, and was ranked the fastest in the world in 1993. The Paragon was aMIMD machine which connected processors via a high speed two-dimensional mesh, allowing processes to execute on separate nodes; communicating via theMessage Passing Interface.[35] By 1995, Cray was also shipping massively parallel systems, e.g. theCray T3E with over 2,000 processors, using a three-dimensionaltorus interconnect.[36][37]
The Paragon architecture soon led to the IntelASCI Red supercomputer in the United States, which held the top supercomputing spot to the end of the 20th century as part of theAdvanced Simulation and Computing Initiative. This was also a mesh-based MIMD massively-parallel system with over 9,000 compute nodes and well over 12 terabytes of disk storage, but used off-the-shelfPentium Pro processors that could be found in everyday personal computers. ASCI Red was the first system ever to break through the 1 teraflop barrier on the MP-Linpack benchmark in 1996; eventually reaching 2 teraflops.[38]

Significant progress was made in the first decade of the 21st century. The efficiency of supercomputers continued to increase, but not dramatically so. TheCray C90 used 500 kilowatts of power in 1991, while by 2003 theASCI Q used 3,000 kW while being 2,000 times faster, increasing the performance per watt 300 fold.[39]
In 2004, theEarth Simulator supercomputer built byNEC at the Japan Agency for Marine-Earth Science and Technology (JAMSTEC) reached 35.9 teraflops, using 640 nodes, each with eight proprietaryvector processors.[40]
TheIBMBlue Gene supercomputer architecture found widespread use in the early part of the 21st century, and 27 of the computers on theTOP500 list used that architecture. The Blue Gene approach is somewhat different in that it trades processor speed for low power consumption so that a larger number of processors can be used at air cooled temperatures. It can use over 60,000 processors, with 2048 processors "per rack", and connects them via a three-dimensional torus interconnect.[41][42]
Progress inChina has been rapid, in that China placed 51st on the TOP500 list in June 2003; this was followed by 14th in November 2003, 10th in June 2004, then 5th during 2005, before gaining the top spot in 2010 with the 2.5 petaflopTianhe-I supercomputer.[43][44]
In July 2011, the 8.1 petaflop JapaneseK computer became the fastest in the world, using over 60,000SPARC64 VIIIfx processors housed in over 600 cabinets. The fact that the K computer is over 60 times faster than the Earth Simulator, and that the Earth Simulator ranks as the 68th system in the world seven years after holding the top spot, demonstrates both the rapid increase in top performance and the widespread growth of supercomputing technology worldwide.[45][46][47] By 2014, the Earth Simulator had dropped off the list and by 2018 the K computer had dropped out of the top 10. By 2018,Summit had become the world's most powerful supercomputer, at 200 petaFLOPS. In 2020, the Japanese once again took the top spot with theFugaku supercomputer, capable of 442 PFLOPS. Finally, starting in 2022 and until the present (as of December 2023[update]), theworld's fastest supercomputer had become the Hewlett Packard EnterpriseFrontier, also known as the OLCF-5 and hosted at theOak Ridge Leadership Computing Facility (OLCF) inTennessee, United States. The Frontier is based on theCray EX, is the world's firstexascalesupercomputer, and uses onlyAMDCPUs andGPUs; it achieved anRmax of 1.102exaFLOPS, which is 1.102 quintillion operations per second.[48][49][50][51][52]
This is a list of the computers which appeared at the top of theTOP500 list since 1993.[53] The "Peak speed" is given as the "Rmax" rating.

TheCoCom and its later replacement, theWassenaar Arrangement, legally regulated, i.e. required licensing and approval and record-keeping; or banned entirely, the export ofhigh-performance computers (HPCs) to certain countries. Such controls have become harder to justify, leading to loosening of these regulations. Some have argued these regulations were never justified.[54][55][56][57][58][59]