Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Tensor Processing Unit

From Wikipedia, the free encyclopedia
AI accelerator ASIC by Google
This article is about the chip developed by Google. For the smartphone system-on-chip, seeGoogle Tensor. For other devices that provide tensor processing for artificial intelligence, seeAI accelerator.
Tensor Processing Unit
Tensor Processing Unit 3.0
DesignerGoogle
Introduced2015[1]
TypeNeural network
Machine learning

Tensor Processing Unit (TPU) is anAI acceleratorapplication-specific integrated circuit (ASIC) developed byGoogle forneural networkmachine learning, using Google's ownTensorFlow software.[2] Google began using TPUs internally in 2015, and in 2018 made them available forthird-party use, both as part of its cloud infrastructure and by offering a smaller version of the chip for sale.

Comparison to CPUs and GPUs

[edit]

Compared to agraphics processing unit, TPUs are designed for a high volume of lowprecision computation (e.g. as little as8-bit precision)[3] with more input/output operations perjoule, without hardware for rasterisation/texture mapping.[4] The TPUASICs are mounted in a heatsink assembly, which can fit in a hard drive slot within a data centerrack, according toNorman Jouppi.[5]

Different types of processors are suited for different types of machine learning models. TPUs are well suited forCNNs, while GPUs have benefits for some fully connected neural networks, and CPUs can have advantages forRNNs.[6]

History

[edit]

According to Jonathan Ross, one of the original TPU engineers,[1] and later the founder ofGroq, three separate groups at Google were developing AI accelerators, with the TPU being the design that was ultimately selected. He was not aware ofsystolic arrays at the time and upon learning the term thought "Oh, that's called a systolic array? It just seemed to make sense."[7]

The tensor processing unit was announced in May 2016 atGoogle I/O, when the company said that the TPU had already been used insidetheir data centers for over a year.[5][4] Google's 2017 paper describing its creation cites previous systolic matrix multipliers of similar architecture built in the 1990s.[8] The chip has been specifically designed for Google'sTensorFlow framework, a symbolic math library which is used formachine learning applications such asneural networks.[9] However, as of 2017 Google still usedCPUs andGPUs for other types ofmachine learning.[5] OtherAI accelerator designs are appearing from other vendors also and are aimed atembedded androbotics markets.

Google's TPUs are proprietary. Some models are commercially available, and on February 12, 2018,The New York Times reported that Google "would allow other companies to buy access to those chips through its cloud-computing service."[10] Google has said that they were used in theAlphaGo versus Lee Sedol series of human-versus-machineGo games,[4] as well as in theAlphaZero system, which producedChess,Shogi and Go playing programs from the game rules alone and went on to beat the leading programs in those games.[11] Google has also used TPUs forGoogle Street View text processing and was able to find all the text in the Street View database in less than five days. InGoogle Photos, an individual TPU can process over 100 million photos a day.[5] It is also used inRankBrain which Google uses to provide search results.[12]

Google provides third parties access to TPUs through itsCloud TPU service as part of theGoogle Cloud Platform[13] and through itsnotebook-based servicesKaggle andColaboratory.[14][15]

Broadcom is a co-developer of TPUs, translating Google's architecture and specifications into manufacturable silicon. It provides proprietary technologies such asSerDes high-speed interfaces, overseeing ASIC design, and managing chip fabrication and packaging through third-party foundries likeTSMC, covering all generations since the program's inception.[16][17][18]

Products

[edit]
Tensor Processing Unit (TPU) generations[19][20][21]
v1v2v3v4[20][22][23]v5e[24]v5p[25][26]v6e (Trillium)[27][28]v7 (Ironwood)[29]
Date introduced20152017201820212023202320242025
Process node28 nm16 nm16 nm7 nmNot listedNot listedNot listedNot listed
Die size (mm2)331< 625< 700< 400300–350Not listedNot listedNot listed
On-chip memory (MiB)283232 (VMEM) + 5 (spMEM)128 (CMEM) + 32 (VMEM) + 10 (spMEM)Not listedNot listedNot listedNot listed
Clock speed (MHz)7007009401050Not listed1750Not listedNot listed
Memory8 GiBDDR316 GiBHBM32 GiB HBM32 GiB HBM16 GB HBM95 GB HBM32 GB192 GB HBM
Memory bandwidth34 GB/s600 GB/s900 GB/s1200 GB/s819 GB/s2765 GB/s1640 GB/s7.37 TB/s
Thermal design power (W)75280220170Not listedNot listedNot listedNot listed
Computational performance (trillion operations per second)2345123275197 (bf16)

393 (int8)

459 (bf16)

918 (int8)

918 (bf16)

1836 (int8)

4614 (fp8)
Energy efficiency (teraOPS/W)0.310.160.561.62Not listedNot listedNot listed4.7


First generation TPU

[edit]

The first-generation TPU is an8-bitmatrix multiplication engine, driven withCISC instructions by the host processor across aPCIe 3.0 bus. It is manufactured on a 28nm process with a die size ≤ 331 mm2. Theclock speed is 700 MHz and it has athermal design power of 28–40 W. It has 28 MiB of on chip memory, and 4 MiB of32-bitaccumulators taking the results of a 256×256systolic array of 8-bitmultipliers.[8] Within the TPU package is 8 GiB ofdual-channel 2133 MHzDDR3 SDRAM offering 34 GB/s of bandwidth.[21] Instructions transfer data to or from the host, perform matrix multiplications orconvolutions, and applyactivation functions.[8]

Second generation TPU

[edit]

The second-generation TPU was announced in May 2017.[30] Google stated the first-generation TPU design was limited bymemory bandwidth and using 16GB ofHigh Bandwidth Memory in the second-generation design increased bandwidth to 600 GB/s and performance to 45 teraFLOPS.[21] The TPUs are then arranged into four-chip modules with a performance of 180 teraFLOPS.[30] Then 64 of these modules are assembled into 256-chip pods with 11.5 petaFLOPS of performance.[30] Notably, while the first-generation TPUs were limited to integers, the second-generation TPUs can also calculate infloating point, introducing thebfloat16 format invented byGoogle Brain. This makes the second-generation TPUs useful for both training and inference of machine learning models. Google has stated these second-generation TPUs will be available on theGoogle Compute Engine for use in TensorFlow applications.[31]

Third generation TPU

[edit]

The third-generation TPU was announced on May 8, 2018.[32] Google announced that processors themselves are twice as powerful as the second-generation TPUs, and would be deployed in pods with four times as many chips as the preceding generation.[33][34] This results in an 8-fold increase in performance per pod (with up to 1,024 chips per pod) compared to the second-generation TPU deployment.

Fourth generation TPU

[edit]

On May 18, 2021, Google CEO Sundar Pichai spoke about TPU v4 Tensor Processing Units during his keynote at the Google I/O virtual conference. TPU v4 improved performance by more than 2x over TPU v3 chips. Pichai said "A single v4 pod contains 4,096 v4 chips, and each pod has 10x the interconnect bandwidth per chip at scale, compared to any other networking technology.”[35] An April 2023 paper by Google claims TPU v4 is 5-87% faster than an NvidiaA100 at machine learningbenchmarks.[36]

There is also an "inference" version, called v4i,[37] that does not requireliquid cooling.[38]

Fifth generation TPU

[edit]

In 2021, Google revealed the physical layout of TPU v5 is being designed with the assistance of a novel application ofdeep reinforcement learning.[39] Google claims TPU v5 is nearly twice as fast as TPU v4,[40] and based on that and the relative performance of TPU v4 over A100, some speculate TPU v5 as being as fast as or faster than anH100.[41]

Similar to the v4i being a lighter-weight version of the v4, the fifth generation has a "cost-efficient"[42] version called v5e.[24] In December 2023, Google announced TPU v5p which is claimed to be competitive with the H100.[43]

Sixth generation TPU

[edit]

In May 2024, at theGoogle I/O conference, Google announced TPU v6, which became available in preview in October 2024.[44] Google claimed a 4.7 times performance increase relative to TPU v5e,[45] via larger matrix multiplication units and an increased clock speed. High bandwidth memory (HBM) capacity and bandwidth have also doubled. A pod can contain up to 256 Trillium units.[46]

Seventh generation TPU

[edit]

In April 2025, at Google Cloud Next conference, Google unveiled TPU v7. This new chip, called Ironwood, will come in two configurations: a 256-chip cluster and a 9,216-chip cluster. Ironwood will have a peak computational performance rate of 4,614 TFLOP/s.[47]

Edge TPU

[edit]

In July 2018, Google announced the Edge TPU. The Edge TPU is Google's purpose-builtASIC chip designed to run machine learning (ML) models foredge computing, meaning it is much smaller and consumes far less power compared to the TPUs hosted in Google datacenters (also known as Cloud TPUs[48]). In January 2019, Google made the Edge TPU available to developers with a line of products under theCoral brand. The Edge TPU is capable of 4 trillion operations per second with 2 W of electrical power.[49]

The product offerings include asingle-board computer (SBC), asystem on module (SoM), aUSB accessory, a miniPCI-e card, and anM.2 card. TheSBC Coral Dev Board and Coral SoM both run Mendel Linux OS – a derivative ofDebian.[50][51] The USB, PCI-e, and M.2 products function as add-ons to existing computer systems, and support Debian-based Linux systems on x86-64 and ARM64 hosts (includingRaspberry Pi).

The machine learning runtime used to execute models on the Edge TPU is based onTensorFlow Lite.[52] The Edge TPU is only capable of accelerating forward-pass operations, which means it's primarily useful for performing inferences (although it is possible to perform lightweight transfer learning on the Edge TPU[53]). The Edge TPU also only supports 8-bit math, meaning that for a network to be compatible with the Edge TPU, it needs to either be trained using the TensorFlow quantization-aware training technique, or since late 2019 it's also possible to use post-training quantization.

On November 12, 2019,Asus announced a pair ofsingle-board computer (SBCs) featuring the Edge TPU. TheAsus Tinker Edge T and Tinker Edge R Board designed forIoT andedgeAI. The SBCs officially supportAndroid andDebianoperating systems.[54][55] ASUS has also demonstrated a mini PC called Asus PN60T featuring the Edge TPU.[56]

On January 2, 2020, Google announced the Coral Accelerator Module and Coral Dev Board Mini, to be demonstrated atCES 2020 later the same month. The Coral Accelerator Module is amulti-chip module featuring the Edge TPU, PCIe and USB interfaces for easier integration. The Coral Dev Board Mini is a smallerSBC featuring the Coral Accelerator Module andMediaTek 8167s SoC.[57][58]

Pixel Neural Core

[edit]
Main article:Pixel Neural Core

On October 15, 2019, Google announced thePixel 4 smartphone, which contains an Edge TPU called thePixel Neural Core. Google describe it as "customized to meet the requirements of key camera features in Pixel 4", using a neural network search that sacrifices some accuracy in favor of minimizing latency and power use.[59]

Google Tensor

[edit]
Main article:Google Tensor

Google followed the Pixel Neural Core by integrating an Edge TPU into a customsystem-on-chip namedGoogle Tensor, which was released in 2021 with thePixel 6 line of smartphones.[60] The Google Tensor SoC demonstrated "extremely large performance advantages over the competition" in machine learning-focused benchmarks; although instantaneous power consumption also was relatively high, the improved performance meant less energy was consumed due to shorter periods requiring peak performance.[61]

Lawsuit

[edit]

In 2019, Singular Computing, founded in 2009 by Joseph Bates, avisiting professor atMIT,[62] filed suit against Google allegingpatent infringement in TPU chips.[63] By 2020, Google had successfully lowered the number of claims the court would consider to just two: claim 53 ofUS 8407273  filed in 2012 and claim 7 ofUS 9218156  filed in 2013, both of which claim adynamic range of 10−6 to 106 for floating point numbers, which the standardfloat16 cannot do (without resorting tosubnormal numbers) as it only has five bits for the exponent. In a 2023 court filing, Singular Computing specifically called out Google's use ofbfloat16, as that exceeds the dynamic range offloat16.[64] Singular claims non-standard floating point formats werenon-obvious in 2009, but Google retorts that the VFLOAT[65] format, with configurable number of exponent bits, existed asprior art in 2002.[66] By January 2024, subsequent lawsuits by Singular had brought the number of patents being litigated up to eight. Towards the end of the trial later that month, Google agreed to a settlement with undisclosed terms.[67][68]

See also

[edit]

References

[edit]
  1. ^abJouppi, Norman; et al. (2017)."In-Datacenter Performance Analysis of a Tensor Processing Unit".Proceedings of the 44th annual international symposium on computer architecture.International Symposium on Computer Architecture.Toronto:Association for Computing Machinery. pp. 1–12.arXiv:1704.04760.doi:10.1145/3079856.3080246.
  2. ^"Cloud Tensor Processing Units (TPUs)".Google Cloud.Archived from the original on 19 June 2022. Retrieved20 July 2020.
  3. ^Armasu, Lucian (2016-05-19)."Google's Big Chip Unveil For Machine Learning: Tensor Processing Unit With 10x Better Efficiency (Updated)".Tom's Hardware. Retrieved2016-06-26.
  4. ^abcJouppi, Norm (May 18, 2016)."Google supercharges machine learning tasks with TPU custom chip".Google Cloud Platform Blog.Archived from the original on 2016-05-18. Retrieved2017-01-22.
  5. ^abcd"Google's Tensor Processing Unit explained: this is what the future of computing looks like".TechRadar.Archived from the original on 2022-02-26. Retrieved2017-01-19.
  6. ^Wang, Yu Emma; Wei, Gu-Yeon; Brooks, David (2019-07-01). "Benchmarking TPU, GPU, and CPU Platforms for Deep Learning".arXiv:1907.10701 [cs.LG].
  7. ^Ross, Jonathan."Jonathan Ross' Post".LinkedIn.
  8. ^abcJouppi, Norman P.; Young, Cliff; Patil, Nishant; Patterson, David; Agrawal, Gaurav; Bajwa, Raminder; Bates, Sarah; Bhatia, Suresh; Boden, Nan; Borchers, Al; Boyle, Rick; Cantin, Pierre-luc; Chao, Clifford; Clark, Chris; Coriell, Jeremy; Daley, Mike; Dau, Matt; Dean, Jeffrey; Gelb, Ben; Ghaemmaghami, Tara Vazir; Gottipati, Rajendra; Gulland, William; Hagmann, Robert; Ho, C. Richard; Hogberg, Doug; Hu, John; Hundt, Robert; Hurt, Dan; Ibarz, Julian; Jaffey, Aaron; Jaworski, Alek; Kaplan, Alexander; Khaitan, Harshit; Koch, Andy; Kumar, Naveen; Lacy, Steve; Laudon, James; Law, James; Le, Diemthu; Leary, Chris; Liu, Zhuyuan; Lucke, Kyle; Lundin, Alan; MacKean, Gordon; Maggiore, Adriana; Mahony, Maire; Miller, Kieran; Nagarajan, Rahul; Narayanaswami, Ravi; Ni, Ray; Nix, Kathy; Norrie, Thomas; Omernick, Mark; Penukonda, Narayana; Phelps, Andy; Ross, Jonathan; Ross, Matt; Salek, Amir; Samadiani, Emad; Severn, Chris; Sizikov, Gregory; Snelham, Matthew; Souter, Jed; Steinberg, Dan; Swing, Andy; Tan, Mercedes; Thorson, Gregory; Tian, Bo; Toma, Horia; Tuttle, Erick; Vasudevan, Vijay; Walter, Richard; Wang, Walter; Wilcox, Eric; Yoon, Doe Hyun (June 26, 2017).In-Datacenter Performance Analysis of a Tensor Processing Unit™. Toronto, Canada.arXiv:1704.04760.
  9. ^"TensorFlow: Open source machine learning"Archived 2016-12-15 at theWayback Machine "It is machine learning software being used for various kinds of perceptual and language understanding tasks" — Jeffrey Dean, minute 0:47 / 2:17 from Youtube clip
  10. ^Metz, Cade (12 February 2018)."Google Makes Its Special A.I. Chips Available to Others".The New York Times.Archived from the original on 2022-04-08. Retrieved2018-02-12.
  11. ^McGourty, Colin (6 December 2017)."DeepMind's AlphaZero crushes chess".chess24.com.Archived from the original on 12 December 2018. Retrieved12 December 2017.
  12. ^"Google's Tensor Processing Unit could advance Moore's Law 7 years into the future".PCWorld.Archived from the original on 2019-02-18. Retrieved2017-01-19.
  13. ^"Frequently Asked Questions | Cloud TPU".Google Cloud. Retrieved2021-01-14.
  14. ^"Google Colaboratory".colab.research.google.com. Retrieved2021-05-15.
  15. ^"Use TPUs | TensorFlow Core".TensorFlow. Retrieved2021-05-15.
  16. ^Tarasov, Katie (2024-08-23)."How Google makes custom chips used to train Apple AI models and its own chatbot, Gemini".CNBC. Retrieved2025-09-19.
  17. ^Gallagher, Dan (2023-09-25)."Broadcom's AI Business Won't Be Easy to Chip Away".The Wall Street Journal. Retrieved2025-09-19.
  18. ^Mann, Tobias (2023-09-23)."FYI: Those fancy 'Google-designed' TPU AI chips had an awful lot of Broadcom help".The Register.
  19. ^Jouppi, Norman P.; Yoon, Doe Hyun; Ashcraft, Matthew; Gottscho, Mark (June 14, 2021).Ten lessons from three generations that shaped Google's TPUv4i(PDF). International Symposium on Computer Architecture. Valencia, Spain.doi:10.1109/ISCA52012.2021.00010.ISBN 978-1-4503-9086-6. Archived fromthe original(PDF) on June 9, 2021. RetrievedJune 18, 2021.
  20. ^ab"System Architecture | Cloud TPU".Google Cloud.Archived from the original on 2022-12-11. Retrieved2022-12-11.
  21. ^abcKennedy, Patrick (22 August 2017)."Case Study on the Google TPU and GDDR5 from Hot Chips 29". Serve The Home.Archived from the original on 23 November 2021. Retrieved23 August 2017.
  22. ^Stay tuned, more information on TPU v4 is coming soonArchived 2023-03-12 at theWayback Machine, retrieved 2020-08-06.
  23. ^Jouppi, Norman P.; Kurian, George; Li, Sheng; Ma, Peter; Nagarajan, Rahul; Nai, Lifeng; Patil, Nishant; Subramanian, Suvinay; Swing, Andy; Towles, Brian; Young, Cliff; Zhou, Xiang; Zhou, Zongwei; Patterson, David (2023). "TPU v4: An Optically Reconfigurable Supercomputer for Machine Learning with Hardware Support for Embeddings".arXiv:2304.01433 [cs.AR].
  24. ^abCloud TPU v5e Inference Public Preview, retrieved 2023-11-06.
  25. ^Cloud TPU v5pArchived 2025-04-28 at theWayback MachineGoogle Cloud. retrieved 2024-04-09
  26. ^Cloud TPU v5p TrainingArchived 2025-02-11 at theWayback Machine, retrieved 2024-04-09.
  27. ^"Introducing Trillium, sixth-generation TPUs".Google Cloud Blog. Retrieved2024-05-29.
  28. ^"TPU v6e".Google Cloud.Archived from the original on 2025-05-10. Retrieved2024-11-10.
  29. ^"Ironwood: The first Google TPU for the age of inference".Google Blog. 9 April 2025.Archived from the original on 2025-05-03. Retrieved2025-04-09.
  30. ^abcBright, Peter (17 May 2017)."Google brings 45 teraflops tensor flow processors to its compute cloud". Ars Technica.Archived from the original on 6 March 2022. Retrieved30 May 2017.
  31. ^Kennedy, Patrick (17 May 2017)."Google Cloud TPU Details Revealed". Serve The Home.Archived from the original on 12 June 2022. Retrieved30 May 2017.
  32. ^Frumusanu, Andre (8 May 2018)."Google I/O Opening Keynote Live-Blog". Archived fromthe original on May 8, 2018. Retrieved9 May 2018.
  33. ^Feldman, Michael (11 May 2018)."Google Offers Glimpse of Third-Generation TPU Processor". Top 500.Archived from the original on 18 November 2021. Retrieved14 May 2018.
  34. ^Teich, Paul (10 May 2018)."Tearing Apart Google's TPU 3.0 AI Coprocessor". The Next Platform.Archived from the original on 15 May 2022. Retrieved14 May 2018.
  35. ^"Google Launches TPU v4 AI Chips".www.hpcwire.com. 20 May 2021.Archived from the original on 7 June 2021. RetrievedJune 7, 2021.
  36. ^Jouppi, Norman (2023-04-20). "TPU v4: An Optically Reconfigurable Supercomputer for Machine Learning with Hardware Support for Embeddings".arXiv:2304.01433 [cs.AR].
  37. ^Kennedy, Patrick (2023-08-29)."Google Details TPUv4 and its Crazy Optically Reconfigurable AI Network".servethehome.com. Retrieved2023-12-16.
  38. ^"Why did Google develop its own TPU chip? In-depth disclosure of team members".censtry.com. 2021-10-20. Retrieved2023-12-16.
  39. ^Mirhoseini, Azalia; Goldie, Anna (2021-06-01)."A graph placement methodology for fast chip design"(PDF).Nature.594 (7962):207–212.doi:10.1038/s41586-022-04657-6.PMID 35361999.S2CID 247855593. Archived fromthe original(PDF) on 2024-08-06. Retrieved2023-06-04.
  40. ^Vahdat, Amin (2023-12-06)."Enabling next-generation AI workloads: Announcing TPU v5p and AI Hypercomputer".Archived from the original on 2024-04-07. Retrieved2024-04-08.
  41. ^Afifi-Sabet, Keumars (2023-12-23)."Google is rapidly turning into a formidable opponent to BFF Nvidia — the TPU v5p AI chip powering its hypercomputer is faster and has more memory and bandwidth than ever before, beating even the mighty H100".TechRadar.Archived from the original on 2025-02-11. Retrieved2024-04-08.
  42. ^"Expanding our AI-optimized infrastructure portfolio: Introducing Cloud TPU v5e and announcing A3 GA". 2023-08-29.Archived from the original on 2023-09-20. Retrieved2023-12-16.
  43. ^"Enabling next-generation AI workloads: Announcing TPU v5p and AI Hypercomputer". 2023-12-06. Retrieved2024-04-09.
  44. ^Lohmeyer, Mark (2024-10-30)."Powerful infrastructure innovations for your AI-first future".
  45. ^Velasco, Alan (2024-05-15)."Google Cloud Unveils Trillium, Its 6th-Gen TPU With A 4.7X AI Performance Leap".HotHardware.HotHardware. Retrieved2024-05-15.
  46. ^"Introducing Trillium, sixth-generation TPUs".Google Cloud Blog. Retrieved2024-05-17.
  47. ^Wiggers, Kyle (2025-04-09)."Ironwood is Google's newest AI accelerator chip".TechCrunch. Retrieved2025-04-10.
  48. ^"Cloud TPU".Google Cloud.Archived from the original on 2025-05-06. Retrieved2021-05-21.
  49. ^"Edge TPU performance benchmarks".Coral. Retrieved2020-01-04.
  50. ^"Dev Board".Coral. Retrieved2021-05-21.
  51. ^"System-on-Module (SoM)".Coral. Retrieved2021-05-21.
  52. ^"Bringing intelligence to the edge with Cloud IoT".Google Blog. 2018-07-25. Retrieved2018-07-25.
  53. ^"Retrain an image classification model on-device".Coral. Retrieved2019-05-03.
  54. ^"組込み総合技術展&IoT総合技術展「ET & IoT Technology 2019」に出展することを発表".Asus.com (in Japanese). Retrieved2019-11-13.
  55. ^Shilov, Anton."ASUS & Google Team Up for 'Tinker Board' AI-Focused Credit-Card Sized Computers".Anandtech.com. Archived fromthe original on November 13, 2019. Retrieved2019-11-13.
  56. ^Aufranc, Jean-Luc (2019-05-29)."ASUS Tinker Edge T & CR1S-CM-A SBC to Feature Google Coral Edge TPU & NXP i.MX 8M Processor".CNX Software - Embedded Systems News. Retrieved2019-11-14.
  57. ^"New Coral products for 2020".Google Developers Blog. Retrieved2020-01-04.
  58. ^"Accelerator Module".Coral. Retrieved2020-01-04.
  59. ^"Introducing the Next Generation of On-Device Vision Models: MobileNetV3 and MobileNetEdgeTPU".Google AI Blog. Retrieved2020-04-16.
  60. ^Gupta, Suyog; White, Marie (November 8, 2021)."Improved On-Device ML on Pixel 6, with Neural Architecture Search".Google AI Blog. Retrieved16 December 2022.
  61. ^Frumusanu, Andrei (November 2, 2021)."Google's Tensor inside of Pixel 6, Pixel 6 Pro: A Look into Performance & Efficiency | Google's IP: Tensor TPU/NPU".AnandTech. Archived fromthe original on November 3, 2021. Retrieved16 December 2022.
  62. ^Hardesty, Larry (2011-01-03)."The surprising usefulness of sloppy arithmetic".MIT. Retrieved2024-01-10.
  63. ^Bray, Hiawatha (2024-01-10)."Local inventor challenges Google in billion-dollar patent fight".Boston Globe.Boston. Archived fromthe original on 2024-01-10. Retrieved2024-01-10.
  64. ^"SINGULAR COMPUTING LLC, Plaintiff, v. GOOGLE LLC, Defendant: Amended Complaint for Patent Infringement"(PDF).rpxcorp.com.RPX Corporation. 2020-03-20. Retrieved2024-01-10.
  65. ^Wang, Xiaojun; Leeser, Miriam (2010-09-01)."VFloat: A Variable Precision Fixed- and Floating-Point Library for Reconfigurable Hardware".ACM Transactions on Reconfigurable Technology and Systems.3 (3):1–34.doi:10.1145/1839480.1839486. Retrieved2024-01-10.
  66. ^"Singular Computing LLC v. Google LLC".casetext.com. 2023-04-06. Archived fromthe original on September 26, 2023. Retrieved2024-01-10.
  67. ^Calkins, Laurel Brubaker (January 24, 2024)."Google Settles AI-Chip Suit That Had Sought Over $5 Billion".Bloomberg Law.
  68. ^Brittain, Blake; Raymond, Ray (January 24, 2024)."Google settles AI-related chip patent lawsuit that sought $1.67 bln".Reuters.


External links

[edit]
Computer
programs
AlphaGo
Versions
Competitions
In popular culture
Other
Machine
learning
Neural networks
Other
Generative
AI
Chatbots
Models
Other
See also
Differentiable computing
General
Hardware
Software libraries
Components
Theory
Design
Applications
Design issues
a subsidiary ofAlphabet
Company
Divisions
Subsidiaries
Active
Defunct
Programs
Events
Infrastructure
People
Current
Former
Criticism
General
Incidents
Other
Software
A–C
D–N
O–Z
Operating systems
Machine learning models
Neural networks
Computer programs
Formats and codecs
Programming languages
Search algorithms
Domain names
Typefaces
A
B
C
D
E
F
G
H
I
J
K
L
M
N
O
P
Q
R
S
T
U
V
W
Y
Hardware
Pixel
Smartphones
Smartwatches
Tablets
Laptops
Other
Nexus
Smartphones
Tablets
Other
Other
Advertising
Antitrust
Intellectual
property
Privacy
Other
Related
Concepts
Products
Android
Street View coverage
YouTube
Other
Documentaries
Books
Popular culture
Other
Retrieved from "https://en.wikipedia.org/w/index.php?title=Tensor_Processing_Unit&oldid=1318477149"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp