Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

History of computing

From Wikipedia, the free encyclopedia

Further information:Timeline of computing
History of computing
Hardware
Software
Computer science
Modern concepts
By country
Timeline of computing
Glossary of computer science

Thehistory of computing extends beyond thehistory of computing hardware andmodern computing technology including earlier methods that relied on pen and paper or chalk and slate, with or without the aid of tables.

Concrete devices

[edit]

Digitalcomputing is intimately tied to the representation ofnumbers.[1] But long beforeabstractions likethe number arose, there were mathematical concepts to serve the purposes of civilization. These concepts are implicit in concrete practices such as:

Numbers

[edit]

Eventually, the concept of numbers became concrete and familiar enough for counting to arise, at times with sing-songmnemonics to teachsequences to others. All known human languages, except thePiraha language, have words for at least thenumerals "one" and "two", and even some animals like theblackbird can distinguish a surprising number of items.[5]

Advances in thenumeral system andmathematical notation eventually led to the discovery of mathematical operations such as addition, subtraction, multiplication, division, squaring, square root, and so forth. Eventually, the operations were formalized, and concepts about the operations became understood well enough to bestated formally, and evenproven. See, for example,Euclid's algorithm for finding the greatest common divisor of two numbers.

By theHigh Middle Ages, thepositionalHindu–Arabic numeral system had reachedEurope, which allowed for the systematic computation of numbers. During this period, the representation of a calculation onpaper allowed the calculation ofmathematical expressions, and the tabulation ofmathematical functions such as thesquare root and thecommon logarithm (for use in multiplication and division), and thetrigonometric functions. By the time ofIsaac Newton's research, paper or vellum was an importantcomputing resource, and even in our present time, researchers likeEnrico Fermi would cover random scraps of paper with calculation, to satisfy their curiosity about an equation.[6] Even into the period of programmable calculators,Richard Feynman would unhesitatingly compute any steps that overflowed thememory of the calculators, by hand, just to learn the answer; by 1976 Feynman had purchased anHP-25 calculator with a 49 program-step capacity; if a differential equation required more than 49 steps to solve, he could just continue his computation by hand.[7]

Early computation

[edit]
For a chronological guide, seeTimeline of computing hardware before 1950.
See also:History of computing hardware

Mathematical statements need not be abstract only; when a statement can be illustrated with actual numbers, the numbers can be communicated and a community can arise. This allows the repeatable, verifiable statements which are the hallmark of mathematics and science. These kinds of statements have existed for thousands of years, and across multiple civilizations, as shown below:

The earliest known tool used for computation is theSumerianabacus, believed to have been invented inBabylonc. 2700–2300 BC. Its original style of usage was by lines drawn in sand with pebbles.[citation needed]

Inc. 1050–771 BC, thesouth-pointing chariot was invented inancient China. It was the first knowngeared mechanism to use adifferential gear, which was later used inanalog computers. TheChinese also invented a more sophisticated abacus from around the 2nd century BC known as theChinese abacus.[citation needed]

In the 3rd century BC,Archimedes used the mechanical principle of balance (seeArchimedes Palimpsest § The Method of Mechanical Theorems) to calculate mathematical problems, such as the number of grains of sand in the universe (The sand reckoner), which also required a recursive notation for numbers (e.g., themyriadmyriad).

TheAntikythera mechanism is believed to be the earliest known geared computing device. It was designed to calculate astronomical positions. It was discovered in 1901 in theAntikythera wreck off the Greek island of Antikythera, between Kythera andCrete, and has been dated tocirca 100 BC.[8]

According toSimon Singh,Muslim mathematicians also made important advances incryptography, such as the development ofcryptanalysis andfrequency analysis byAlkindus.[9][10]Programmable machines were also invented byMuslim engineers, such as the automaticflute player by theBanū Mūsā brothers.[11]

During the Middle Ages, several European philosophers made attempts to produce analog computer devices. Influenced by the Arabs andScholasticism, Majorcan philosopherRamon Llull (1232–1315) devoted a great part of his life to defining and designing severallogical machines that, by combining simple and undeniable philosophical truths, could produce all possible knowledge. These machines were never actually built, as they were more of athought experiment to produce new knowledge in systematic ways; although they could make simple logical operations, they still needed a human being for the interpretation of results. Moreover, they lacked a versatile architecture, each machine serving only very concrete purposes. Despite this, Llull's work had a strong influence onGottfried Leibniz (early 18th century), who developed his ideas further and built several calculating tools using them.

The apex of this early era of mechanical computing can be seen in theDifference Engine and its successor theAnalytical Engine both byCharles Babbage. Babbage never completed constructing either engine, but in 2002Doron Swade and a group of other engineers at theScience Museum in London completed Babbage's Difference Engine using only materials that would have been available in the 1840s.[12] By following Babbage's detailed design they were able to build a functioning engine, allowing historians to say, with some confidence, that if Babbage had been able to complete his Difference Engine it would have worked.[13] The additionally advanced Analytical Engine combined concepts from his previous work and that of others to create a device that, if constructed as designed, would have possessed many properties of a modern electronic computer, such as an internal "scratch memory" equivalent toRAM, multiple forms of output including a bell, a graph-plotter, and simple printer, and a programmable input-output "hard" memory ofpunch cards which it could modify as well as read. The key advancement that Babbage's devices possessed beyond those created before him was that each component of the device was independent of the rest of the machine, much like the components of a modern electronic computer. This was a fundamental shift in thought; previous computational devices served only a single purpose but had to be at best disassembled and reconfigured to solve a new problem. Babbage's devices could be reprogrammed to solve new problems by the entry of new data and act upon previous calculations within the same series of instructions.Ada Lovelace took this concept one step further, by creating a program for the Analytical Engine to calculateBernoulli numbers, a complex calculation requiring a recursive algorithm. This is considered to be the first example of a true computer program, a series of instructions that act upon data not known in full until the program is run.

Following Babbage, although unaware of his earlier work,Percy Ludgate[14][15] in 1909 published the 2nd of the only two designs for mechanical analytical engines in history.[16] Two other inventors,Leonardo Torres Quevedo[17] andVannevar Bush,[18] also did follow-on research based on Babbage's work. In hisEssays on Automatics (1914) Torres presented the design of an electromechanical calculating machine and introduced the idea ofFloating-point arithmetic.[19][20] In 1920, to celebrate the 100th anniversary of the invention of thearithmometer, Torres presented in Paris theElectromechanical Arithmometer, an arithmetic unit connected to a remote typewriter, on which commands could be typed and the results printed automatically.[21][22] Bush's paperInstrumental Analysis (1936) discussed using existing IBM punch card machines to implement Babbage's design. In the same year, he started the Rapid Arithmetical Machine project to investigate the problems of constructing an electronic digital computer.

Several examples of analog computation survived into recent times. Aplanimeter is a device that does integrals, usingdistance as the analog quantity. Until the 1980s,HVAC systems usedair both as the analog quantity and the controlling element. Unlike modern digital computers, analog computers are not very flexible and need to be reconfigured (i.e., reprogrammed) manually to switch them from working on one problem to another. Analog computers had an advantage over early digital computers in that they could be used to solve complex problems using behavioral analogues while the earliest attempts at digital computers were quite limited.

ASmith Chart is a well-knownnomogram.

Since computers were rare in this era, the solutions were oftenhard-coded into paper forms such asnomograms,[23] which could then produce analog solutions to these problems, such as the distribution of pressures and temperatures in a heating system.

Digital electronic computers

[edit]
Main articles:Modern computers,History of computing hardware, andHistory of computing hardware (1960s–present)

The "brain" [computer] may one day come down to our level [of the common people] and help with our income-tax and book-keeping calculations. But this is speculation and there is no sign of it so far.

— British newspaperThe Star in a June 1949 news article about theEDSAC computer, long before the era of the personal computers.[24]

In an 1886 letter,Charles Sanders Peirce described how logical operations could be carried out by electrical switching circuits.[25] During 1880-81 he showed thatNOR gates alone (orNAND gates alone) can be used to reproduce the functions of all the otherlogic gates, but this work on it was unpublished until 1933.[26] The first published proof was byHenry M. Sheffer in 1913, so the NAND logical operation is sometimes calledSheffer stroke; thelogical NOR is sometimes calledPeirce's arrow.[27] Consequently, these gates are sometimes calleduniversal logic gates.[28]

Eventually,vacuum tubes replaced relays for logic operations.Lee De Forest's modification, in 1907, of theFleming valve can be used as a logic gate.Ludwig Wittgenstein introduced a version of the 16-rowtruth table as proposition 5.101 ofTractatus Logico-Philosophicus (1921).Walther Bothe, inventor of thecoincidence circuit, got part of the 1954Nobel Prize in physics, for the first modern electronic AND gate in 1924.Konrad Zuse designed and built electromechanical logic gates for his computerZ1 (from 1935 to 1938).

The first recorded idea of usingdigital electronics for computing was the 1931 paper "The Use of Thyratrons for High Speed Automatic Counting of Physical Phenomena" byC. E. Wynn-Williams.[29] From 1934 to 1936,NEC engineerAkira Nakashima,Claude Shannon, andVictor Shestakov published papers introducingswitching circuit theory, using digital electronics forBoolean algebraic operations.[30][31][32][33]

In 1936Alan Turing published his seminal paperOn Computable Numbers, with an Application to the Entscheidungsproblem[34] in which he modeled computation in terms of a one-dimensional storage tape, leading to the idea of theUniversal Turing machine andTuring-complete systems.[citation needed]

The first digital electronic computer was developed between April 1936 and June 1939 at the IBM Patent Department in Endicott, New York, by Arthur Halsey Dickinson.[35][36][37] In this computer IBM introduced, a calculating device with a keyboard, processor and electronic output (display). The competitor to IBM was the digital electronic computer NCR3566, developed in NCR, Dayton, Ohio by Joseph Desch and Robert Mumma in the period April 1939 - August 1939.[38][39] The IBM and NCR machines were decimal, executing addition and subtraction in binary position code.

In December 1939John Atanasoff andClifford Berry completed their experimental model to prove the concept of theAtanasoff–Berry computer (ABC) which began development in 1937.[40] This experimental model is binary, executed addition and subtraction in octal binary code and is the first binary digitalelectronic computing device. The Atanasoff–Berry computer was intended to solve systems of linear equations, though it was not programmable. The computer was never truly completed due to Atanasoff's departure fromIowa State University in 1942 to work for the United States Navy.[41][42] Many people credit ABC with many of the ideas used in later developments during the age of early electronic computing.[43]

TheZ3 computer, built byGerman inventorKonrad Zuse in 1941, was the first programmable, fully automatic computing machine, but it was not electronic.

During World War II, ballistics computing was done by women, who were hired as "computers." The term computer remained one that referred to mostly women (now seen as "operator") until 1945, after which it took on the modern definition of machinery it presently holds.[44]

TheENIAC (Electronic Numerical Integrator And Computer) was the first electronic general-purpose computer, announced to the public in 1946. It was Turing-complete,[45] digital, and capable of being reprogrammed to solve a full range of computing problems. Women implemented the programming for machines like the ENIAC, and men created the hardware.[44]

TheManchester Baby was the first electronicstored-program computer. It was built at theVictoria University of Manchester byFrederic C. Williams,Tom Kilburn andGeoff Tootill, and ran its first program on 21 June 1948.[46]

William Shockley,John Bardeen andWalter Brattain atBell Labs invented the first workingtransistor, thepoint-contact transistor, in 1947, followed by thebipolar junction transistor in 1948.[47][48] At theUniversity of Manchester in 1953, a team under the leadership ofTom Kilburn designed and built the firsttransistorized computer, called theTransistor Computer, a machine using the newly developed transistors instead of valves.[49] The first stored-program transistor computer was the ETL Mark III, developed by Japan's Electrotechnical Laboratory[50][51][52] from 1954[53] to 1956.[51] However, early junction transistors were relatively bulky devices that were difficult to manufacture on amass-production basis, which limited them to a number of specialized applications.[54]

In 1954, 95% of computers in service were being used for engineering and scientific purposes.[55]

Personal computers

[edit]

Themetal–oxide–silicon field-effect transistor (MOSFET), also known as the MOS transistor, was invented at Bell Labs between 1955 and 1960,[56][57][58][59][60][61][62] It was the first truly compact transistor that could beminiaturised andmass-produced for a wide range of uses.[54] The MOSFET made it possible to buildhigh-densityintegrated circuit chips.[63][64] The MOSFET is the most widely used transistor in computers,[65][66] and is the fundamental building block ofdigital electronics.[67]

Thesilicon-gate MOS integrated circuit was developed byFederico Faggin atFairchild Semiconductor in 1968.[68] This led to the development of the first single-chipmicroprocessor, theIntel 4004.[69] The Intel 4004 was developed as a single-chip microprocessor from 1969 to 1970, led by Intel's Federico Faggin,Marcian Hoff, andStanley Mazor, and Busicom's Masatoshi Shima.[70] The chip was mainly designed and realized by Faggin, with his silicon-gate MOS technology.[69] The microprocessor led to the microcomputer revolution, with the development of themicrocomputer, which would later be called thepersonal computer (PC).

Most early microprocessors, such as theIntel 8008 andIntel 8080, were8-bit. Texas Instruments released the first fully16-bit microprocessor, theTMS9900 processor, in June 1976.[71] They used the microprocessor in the TI-99/4 andTI-99/4A computers.

The 1980s brought about significant advances with microprocessors that greatly impacted the fields of engineering and other sciences. TheMotorola 68000 microprocessor had a processing speed that was far superior to the other microprocessors being used at the time. Because of this, having a newer, faster microprocessor allowed for the newermicrocomputers that came along after to be more efficient in the amount of computing they were able to do. This was evident in the 1983 release of theApple Lisa. The Lisa was one of the first personal computers with agraphical user interface (GUI) that was sold commercially. It ran on the Motorola 68000 CPU and used both dual floppy disk drives and a 5 MB hard drive for storage. The machine also had 1MB ofRAM used for running software from disk without rereading the disk persistently.[72] After the failure of the Lisa in terms of sales, Apple released itsfirst Macintosh computer, still running on the Motorola 68000 microprocessor, but with only 128KB of RAM, one floppy drive, and no hard drive to lower the price.

In the late 1980s and early 1990s, computers became more useful for personal and work purposes, such asword processing.[73] In 1989, Apple released theMacintosh Portable, it weighed 7.3 kg (16 lb) and was extremely expensive, costing US$7,300. At launch, it was one of the most powerful laptops available, but due to the price and weight, it was not met with great success and was discontinued only two years later. That same year Intel introduced the Touchstone Deltasupercomputer, which had 512 microprocessors. This technological advancement was very significant, as it was used as a model for some of the fastest multi-processor systems in the world. It was even used as a prototype for Caltech researchers, who used the model for projects like real-time processing of satellite images and simulating molecular models for various fields of research.

Supercomputers

[edit]

In terms of supercomputing, the first widely acknowledged supercomputer was theControl Data Corporation (CDC)6600[74] built in 1964 bySeymour Cray. Its maximum speed was 40  MHz or 3 million floating point operations per second (FLOPS). The CDC 6600 was replaced by theCDC 7600 in 1969;[75] although its normal clock speed was not faster than the 6600, the 7600 was still faster due to its peak clock speed, which was approximately 30 times faster than that of the 6600. Although CDC was a leader in supercomputers, their relationship with Seymour Cray (which had already been deteriorating) completely collapsed. In 1972, Cray left CDC and began his own company,Cray Research Inc.[76] With support from investors in Wall Street, an industry fueled by the Cold War, and without the restrictions he had within CDC, he created theCray-1 supercomputer. With a clock speed of 80  MHz or 136 megaFLOPS, Cray developed a name for himself in the computing world. By 1982, Cray Research produced theCray X-MP equipped with multiprocessing and in 1985 released theCray-2, which continued with the trend of multiprocessing and clocked at 1.9 gigaFLOPS. Cray Research developed theCray Y-MP in 1988, however afterward struggled to continue to produce supercomputers. This was largely because the Cold War had ended, and the demand for cutting-edge computing by colleges and the government declined drastically and the demand for microprocessing units increased.

In 1998,David Bader developed the firstLinux supercomputer using commodity parts.[77] While at the University of New Mexico, Bader sought to build a supercomputer running Linux using consumer off-the-shelf parts and a high-speed low-latency interconnection network. The prototype utilized an Alta Technologies "AltaCluster" of eight dual, 333  MHz, Intel Pentium II computers running a modified Linux kernel. Bader ported a significant amount of software to provide Linux support for necessary components as well as code from members of the National Computational Science Alliance (NCSA) to ensure interoperability, as none of it had been run on Linux previously.[78] Using the successful prototype design, he led the development of "RoadRunner," the first Linux supercomputer for open use by the national science and engineering community via the National Science Foundation's National Technology Grid. RoadRunner was put into production use in April 1999. At the time of its deployment, it was considered one of the 100 fastest supercomputers in the world.[78][79] Though Linux-based clusters using consumer-grade parts, such asBeowulf, existed before the development of Bader's prototype and RoadRunner, they lacked the scalability, bandwidth, andparallel computing capabilities to be considered "true" supercomputers.[78]

Today, supercomputers are still used by the governments of the world and educational institutions for computations such as simulations of natural disasters, genetic variant searches within a population relating to disease, and more. As of November 2024[update], the fastest supercomputer isEl Capitan.

Navigation and astronomy

[edit]

Starting with known special cases, the calculation of logarithms and trigonometric functions can be performed by looking up numbers in amathematical table, andinterpolating between known cases. For small enough differences, this linear operation was accurate enough for use innavigation andastronomy in theAge of Exploration. The uses of interpolation have thrived in the past 500 years: by the twentieth centuryLeslie Comrie andW.J. Eckert systematized the use of interpolation in tables of numbers for punch card calculation.

Weather prediction

[edit]

The numerical solution of differential equations, notably theNavier-Stokes equations was an important stimulus to computing,withLewis Fry Richardson's numerical approach to solving differential equations. The first computerized weather forecast was performed in 1950 by a team composed of American meteorologistsJule Charney,Philip Duncan Thompson, Larry Gates, and Norwegian meteorologistRagnar Fjørtoft, applied mathematicianJohn von Neumann, andENIAC programmerKlara Dan von Neumann.[80][81][82] To this day, some of the most powerful computer systems on Earth are used forweather forecasts.[83]

Symbolic computations

[edit]
Main article:Computer algebra

By the late 1960s, computer systems could performsymbolic algebraic manipulations well enough to pass college-levelcalculus courses.[citation needed]

Important women and their contributions

[edit]

Women are often underrepresented inSTEM fields when compared to their male counterparts.[84] In the modern era before the 1960s, computing was widely seen as "women's work" since it was associated with the operation oftabulating machines and other mechanical office work.[85][86] The accuracy of this association varied from place to place. In America,Margaret Hamilton recalled an environment dominated by men,[87] whileElsie Shutt recalled surprise at seeing even half of the computer operators at Raytheon were men.[88] Machine operators in Britain were mostly women into the early 1970s.[89] As these perceptions changed and computing became a high-status career, the field became more dominated by men.[90][91][92] ProfessorJanet Abbate, in her bookRecoding Gender, writes:

Yet women were a significant presence in the early decades of computing. They made up the majority of the first computer programmers during World War II; they held positions of responsibility and influence in the early computer industry; and they were employed in numbers that, while a small minority of the total, compared favorably with women's representation in many other areas of science and engineering. Some female programmers of the 1950s and 1960s would have scoffed at the notion that programming would ever be considered a masculine occupation, yet these women’s experiences and contributions were forgotten all too quickly.[93]

Some notable examples of women in the history of computing are:

See also

[edit]

References

[edit]
  1. ^"Digital Computing - Dictionary definition of Digital Computing | Encyclopedia.com: FREE online dictionary".www.encyclopedia.com. Retrieved2017-09-11.
  2. ^"One-to-One Correspondence: 0.5".Victoria Department of Education and Early Childhood Development. Archived fromthe original on 20 November 2012.
  3. ^Ifrah, Georges (2000),The Universal History of Numbers: From prehistory to the invention of the computer.,John Wiley and Sons, p. 48,ISBN 0-471-39340-1
  4. ^W., Weisstein, Eric."3, 4, 5 Triangle".mathworld.wolfram.com. Retrieved2017-09-11.{{cite web}}: CS1 maint: multiple names: authors list (link)
  5. ^Lorenz, Konrad (1961).King Solomon's Ring. Translated by Marjorie Kerr Wilson. London: Methuen.ISBN 0-416-53860-6.{{cite book}}:ISBN / Date incompatibility (help)
  6. ^"DIY: Enrico Fermi's Back of the Envelope Calculations".
  7. ^"Try numbers" was one of Feynman's problem solving techniques.
  8. ^Ouellette, Jennifer (12 March 2021)."Scientists solve another piece of the puzzling Antikythera mechanism". Ars Technica.
  9. ^Simon Singh,The Code Book, pp. 14-20
  10. ^"Al-Kindi, Cryptgraphy, Codebreaking and Ciphers". 9 June 2003. Retrieved2022-07-03.
  11. ^Koetsier, Teun (2001), "On the prehistory of programmable machines: musical automata, looms, calculators",Mechanism and Machine Theory,36 (5), Elsevier:589–603,doi:10.1016/S0094-114X(01)00005-2.
  12. ^"A 19th-Century Mathematician Finally Proves Himself".NPR.org. Retrieved2022-10-24.
  13. ^"A Modern Sequel | Babbage Engine | Computer History Museum".www.computerhistory.org. Retrieved2022-10-24.
  14. ^Randell 1982, pp. 4–5.
  15. ^"Percy Ludgate's Analytical Machine".fano.co.uk. Retrieved29 October 2018.
  16. ^"Percy E. Ludgate Prize in Computer Science"(PDF).The John Gabriel Byrne Computer Science Collection. Retrieved2020-01-15.
  17. ^Randell 1982, pp. 6, 11–13.
  18. ^Randell 1982, pp. 13, 16–17.
  19. ^L. Torres Quevedo (1914). "Ensayos sobre Automática – Su definicion. Extension teórica de sus aplicaciones".Revista de la Academia de Ciencias Exacta, Revista 12:391–418.
  20. ^Torres Quevedo, L. (1915)."Essais sur l'Automatique - Sa définition. Etendue théorique de ses applications".Revue Génerale des Sciences Pures et Appliquées.2:601–611.
  21. ^"Computer Pioneers by J.A.N. Lee - Leonardo Torres Y Quevedo". Retrieved3 February 2018.
  22. ^Bromley 1990.
  23. ^Steinhaus, H. (1999).Mathematical Snapshots (3rd ed.). New York: Dover. pp. 92–95, 301.
  24. ^"Tutorial Guide to the EDSAC Simulator"(PDF). Retrieved2020-01-15.
  25. ^Peirce, C. S., "Letter, Peirce toA. Marquand", dated 1886,Writings of Charles S. Peirce, v. 5, 1993, pp. 421–423. SeeBurks, Arthur W. (September 1978)."Review: Charles S. Peirce,The new elements of mathematics".Bulletin of the American Mathematical Society.84 (5): 917.doi:10.1090/S0002-9904-1978-14533-9.
  26. ^Peirce, C. S. (manuscript winter of 1880–81), "A Boolian Algebra with One Constant", published 1933 inCollected Papers v. 4, paragraphs 12–20. Reprinted 1989 inWritings of Charles S. Peirce v. 4,pp. 218–212. See Roberts, Don D. (2009),The Existential Graphs of Charles S. Peirce, p. 131.
  27. ^Büning, Hans Kleine; Lettmann, Theodor (1999).Propositional logic: deduction and algorithms. Cambridge University Press. p. 2.ISBN 978-0-521-63017-7.
  28. ^Bird, John (2007).Engineering mathematics. Newnes. p. 532.ISBN 978-0-7506-8555-9.
  29. ^Wynn-Williams, C. E. (July 2, 1931), "The Use of Thyratrons for High Speed Automatic Counting of Physical Phenomena",Proceedings of the Royal Society A,132 (819):295–310,Bibcode:1931RSPSA.132..295W,doi:10.1098/rspa.1931.0102
  30. ^Yamada, Akihiko (2004)."History of Research on Switching Theory in Japan".IEEJ Transactions on Fundamentals and Materials.124 (8).Institute of Electrical Engineers of Japan:720–726.Bibcode:2004IJTFM.124..720Y.doi:10.1541/ieejfms.124.720.
  31. ^"Switching Theory/Relay Circuit Network Theory/Theory of Logical Mathematics".IPSJ Computer Museum,Information Processing Society of Japan.
  32. ^Stanković, Radomir S.[in German];Astola, Jaakko Tapio[in Finnish], eds. (2008).Reprints from the Early Days of Information Sciences: TICSP Series On the Contributions of Akira Nakashima to Switching Theory(PDF). Tampere International Center for Signal Processing (TICSP) Series. Vol. 40.Tampere University of Technology, Tampere, Finland.ISBN 978-952-15-1980-2.ISSN 1456-2774. Archived fromthe original(PDF) on 2021-03-08.{{cite book}}: CS1 maint: location missing publisher (link) (3+207+1 pages)10:00 min
  33. ^Stanković, Radomir S.; Astola, Jaakko T.; Karpovsky, Mark G."Some Historical Remarks on Switching Theory"(PDF). Tampere International Center for Signal Processing,Tampere University of Technology.CiteSeerX 10.1.1.66.1248. Archived fromthe original(PDF) on 5 July 2017.
  34. ^*Turing, Alan M. (1936), "On Computable Numbers, with an Application to the Entscheidungsproblem",Proceedings of the London Mathematical Society, 2, vol. 42 (published 1937), pp. 230–265,doi:10.1112/plms/s2-42.1.230,S2CID 73712 (andTuring, Alan M. (1938), "On Computable Numbers, with an Application to the Entscheidungsproblem. A correction",Proceedings of the London Mathematical Society, 2, vol. 43, no. 6 (published 1937), pp. 544–546,doi:10.1112/plms/s2-43.6.544)
  35. ^Dickinson, A.H., "Accounting Apparatus",U.S. patent 2,580,740, filed Jan. 20, 1940, granted Jan. 1, 1952,
  36. ^Pugh, Emerson W. (1996).Building IBM: Shaping an Industry and its Technology.The MIT Press.
  37. ^"Patents and Inventions".IBM100. 7 March 2012.
  38. ^Desch, J.R., "Calculating Machine",U.S. patent 2,595,045, filed March 20, 1940, granted Apr. 29, 1952,
  39. ^"Interview with Robert E. Mumma" (Interview). Interviewed by William Aspray. Dayton, OH, Charles Babbage Institute, Center for the History of Information Processing. 19 April 1984.
  40. ^Larson E., "Findings of Fact, Conclusions of Law and Order for Judgement", US District Court, District of Minnesota, Fourth Division, 19 Oct. 1973, ushistory.org/more/eniac/index.htm, ushistory.org/more/eniac/intro.htm
  41. ^"Atanasoff-Berry Computer Operation/Purpose".
  42. ^"interview with John V. Atanasoff"(PDF) (Interview). Interviewed by Tropp H.S. Computer Oral History Collection, 1969-1973, 1979, Smithsonian National Museum of American History, Lemelson Center for the Study of Invention and Innovation. May 11, 1972. Archived fromthe original(PDF) on January 29, 2022. RetrievedJanuary 9, 2022.
  43. ^"Atanasoff-Berry Computer Court Case".
  44. ^abLight, Jennifer S. (July 1999). "When Computers Were Women".Technology and Culture.40 (3):455–483.doi:10.1353/tech.1999.0128.S2CID 108407884.
  45. ^rudd."Early Turing-complete Computers | Rudd Canaday". Retrieved2024-04-17.
  46. ^Enticknap, Nicholas (Summer 1998)."Computing's Golden Jubilee".Resurrection (20). The Computer Conservation Society.ISSN 0958-7403.
  47. ^Lee, Thomas H. (2003).The Design of CMOS Radio-Frequency Integrated Circuits(PDF).Cambridge University Press.ISBN 9781139643771. Archived fromthe original(PDF) on 2019-12-09. Retrieved2019-09-16.
  48. ^Puers, Robert; Baldi, Livio; Voorde, Marcel Van de; Nooten, Sebastiaan E. van (2017).Nanoelectronics: Materials, Devices, Applications, 2 Volumes.John Wiley & Sons. p. 14.ISBN 9783527340538.
  49. ^Lavington, Simon (1998),A History of Manchester Computers (2 ed.), Swindon: The British Computer Society, pp. 34–35
  50. ^"Early Computers".Information Processing Society of Japan.
  51. ^ab"Electrotechnical Laboratory ETL Mark III Transistor-Based Computer".Information Processing Society of Japan.
  52. ^"Early Computers: Brief History".Information Processing Society of Japan.
  53. ^Fransman, Martin (1993).The Market and Beyond: Cooperation and Competition in Information Technology.Cambridge University Press. p. 19.ISBN 9780521435253.
  54. ^abMoskowitz, Sanford L. (2016).Advanced Materials Innovation: Managing Global Technology in the 21st century.John Wiley & Sons. pp. 165–167.ISBN 9780470508923.
  55. ^Ensmenger, Nathan (2010).The Computer Boys Take Over. MIT Press. p. 58.ISBN 978-0-262-05093-7.
  56. ^US2802760A, Lincoln, Derick & Frosch, Carl J., "Oxidation of semiconductive surfaces for controlled diffusion", issued 1957-08-13 
  57. ^Huff, Howard; Riordan, Michael (2007-09-01)."Frosch and Derick: Fifty Years Later (Foreword)".The Electrochemical Society Interface.16 (3): 29.doi:10.1149/2.F02073IF.ISSN 1064-8208.
  58. ^Frosch, C. J.; Derick, L (1957)."Surface Protection and Selective Masking during Diffusion in Silicon".Journal of the Electrochemical Society.104 (9): 547.doi:10.1149/1.2428650.
  59. ^KAHNG, D. (1961)."Silicon-Silicon Dioxide Surface Device".Technical Memorandum of Bell Laboratories:583–596.doi:10.1142/9789814503464_0076.ISBN 978-981-02-0209-5.{{cite journal}}:ISBN / Date incompatibility (help)
  60. ^Lojek, Bo (2007).History of Semiconductor Engineering. Berlin, Heidelberg: Springer-Verlag Berlin Heidelberg. p. 321.ISBN 978-3-540-34258-8.
  61. ^Ligenza, J.R.; Spitzer, W.G. (1960)."The mechanisms for silicon oxidation in steam and oxygen".Journal of Physics and Chemistry of Solids.14:131–136.Bibcode:1960JPCS...14..131L.doi:10.1016/0022-3697(60)90219-5.
  62. ^Lojek, Bo (2007).History of Semiconductor Engineering.Springer Science & Business Media. p. 120.ISBN 9783540342588.
  63. ^"Who Invented the Transistor?".Computer History Museum. 4 December 2013. Retrieved20 July 2019.
  64. ^Hittinger, William C. (1973). "Metal-Oxide-Semiconductor Technology".Scientific American.229 (2):48–59.Bibcode:1973SciAm.229b..48H.doi:10.1038/scientificamerican0873-48.ISSN 0036-8733.JSTOR 24923169.
  65. ^"Dawon Kahng".National Inventors Hall of Fame. Retrieved27 June 2019.
  66. ^"Martin Atalla in Inventors Hall of Fame, 2009". Retrieved21 June 2013.
  67. ^"Triumph of the MOS Transistor".YouTube.Computer History Museum. 6 August 2010.Archived from the original on 2021-12-21. Retrieved21 July 2019.
  68. ^"1968: Silicon Gate Technology Developed for ICs".Computer History Museum. Retrieved22 July 2019.
  69. ^ab"1971: Microprocessor Integrates CPU Function onto a Single Chip".Computer History Museum. Retrieved22 July 2019.
  70. ^Faggin, Federico (Winter 2009). "The Making of the First Microprocessor".IEEE Solid-State Circuits Magazine.1 (1):8–21.doi:10.1109/MSSC.2008.930938.S2CID 46218043.
  71. ^Conner, Stuart."Stuart's TM 990 Series 16-bit Microcomputer Modules".www.stuartconner.me.uk. Retrieved2017-09-05.
  72. ^"Computers | Timeline of Computer History | Computer History Museum".www.computerhistory.org. Retrieved2022-12-26.
  73. ^"A brave new world: the 1980s home computer boom".HistoryExtra. Retrieved2024-04-18.
  74. ^Vaughan-Nichols, Steven (November 27, 2017)."A super-fast history of supercomputers: From the CDC 6600 to the Sunway TaihuLight".
  75. ^"CDC 7600".
  76. ^"Seymour R. Cray".Encyclopædia Britannica.
  77. ^"David Bader Selected to Receive the 2021 IEEE Computer Society Sidney Fernbach Award". IEEE Computer Society. September 22, 2021. Retrieved2023-10-12.
  78. ^abcBader, David A. (2021)."Linux and Supercomputing: How My Passion for Building COTS Systems Led to an HPC Revolution".IEEE Annals of the History of Computing.43 (3):73–80.doi:10.1109/MAHC.2021.3101415.S2CID 237318907.
  79. ^Fleck, John (April 8, 1999). "UNM to crank up $400,000 supercomputer today".Albuquerque Journal. p. D1.
  80. ^Charney, J.G.;Fjörtoft, R.;von Neumann, J. (November 1950)."Numerical Integration of the Barotropic Vorticity Equation".Tellus.2 (4):237–254.Bibcode:1950Tell....2..237C.doi:10.3402/tellusa.v2i4.8607.
  81. ^Witman, Sarah (16 June 2017)."Meet the Computer Scientist You Should Thank For Your Smartphone's Weather App".Smithsonian. Retrieved22 July 2017.
  82. ^Edwards, Paul N. (2010).A Vast Machine: Computer Models, Climate Data, and the Politics of Global Warming. The MIT Press.ISBN 978-0262013925. Retrieved2020-01-15.
  83. ^US Department of Commerce, NOAA."About Supercomputers".www.weather.gov. Retrieved2024-04-18.
  84. ^Myers, Blanca (March 3, 2018)."Women and Minorities in Tech, By the Numbers".Wired.
  85. ^Ensmenger, Nathan (2012).The Computer Boys Take Over. MIT Press. p. 38.ISBN 978-0-262-51796-6.
  86. ^Hicks, Mar (2017).Programmed inequality: how Britain discarded women technologists and lost its edge in computing. MIT Press. p. 1.ISBN 978-0-262-53518-2.OCLC 1089728009.
  87. ^Creighton, Jolene (July 7, 2016)."Margaret Hamilton: The Untold Story of the Woman Who Took Us to the Moon". Futurism.com.
  88. ^Thompson, Clive (February 13, 2019)."The Secret History of Women in Coding".New York Times.
  89. ^Hicks 2017, pp. 215–216: "The Civil Service's computing workforce continued to bifurcate along both gendered and class lines, even though among machine operators in industry and government there were still more than 6.5 times as many women as men in 1971."
  90. ^Cohen, Rhaina (September 7, 2016)."What Programming's Past Reveals About Today's Gender Pay Gap".The Atlantic.
  91. ^Hicks 2017, pp. 1–9: "In the 1940s, computer operation and programming was viewed as women's work—but by the 1960s, as computing gained prominence and influence, men displaced the thousands of women who had been pioneers in a feminized field of endeavor, and the field acquired a distinctly masculine image ... Soon, women became synonymous with office machine operators and their work became tied to typewriters, desktop accounting machines, and room-sized punched card equipment installations ... Their alignment with machine work in offices persisted through waves of equipment upgrades and eventually through the changeover from electromechanical to electronic systems."
  92. ^Ensmenger 2012, p. 239: "Over the 1960s, developments in the computing professions were creating new barriers to female participation. An activity originally intended to be performed by low-status, clerical—and more often than not, female—computer programming was gradually and deliberately transformed into a high-status, scientific, and masculine discipline .... In 1965, for example, the Association for Computing Machinery imposed a four-year degree requirement for membership that, in an era when there were almost twice as many male as there were female college undergraduates, excluded significantly more women than men ... Similarly, certification programs or licensing requirements erected barriers to entry that disproportionately affected women."
  93. ^Abbate, Janet (2012).Recoding gender : women's changing participation in computing. Cambridge, Mass.: MIT Press. p. 1.ISBN 978-0-262-30546-4.OCLC 813929041.

Works cited

[edit]

Further reading

[edit]

External links

[edit]

British history links

[edit]
Retrieved from "https://en.wikipedia.org/w/index.php?title=History_of_computing&oldid=1319455259"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp