Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Nvidia DGX

From Wikipedia, the free encyclopedia
Line of Nvidia produced servers and workstations
This articlecontainspromotional content. Please helpimprove it by removingpromotional language and inappropriateexternal links, and by adding encyclopedic text written from aneutral point of view.
See ouradvice if the article is about you and read ourscam warning in case someone asks for money to edit this article.
(January 2024) (Learn how and when to remove this message)

DGX
A rack containing five DGX-1 supercomputers
ManufacturerNvidia
Released2016 (2016)

TheNvidia DGX (Deep GPU Xceleration) is a series ofservers andworkstations designed byNvidia, primarily geared towards enhancingdeep learning applications through the use ofgeneral-purpose computing on graphics processing units (GPGPU). These systems typically come in a rackmount format featuring high-performancex86 server CPUs on the motherboard.

The core feature of a DGX system is its inclusion of 4 to 8Nvidia Tesla GPU modules, which are housed on an independent system board. These GPUs can be connected either via a version of theSXM socket or aPCIe x16 slot, facilitating flexible integration within the system architecture. To manage the substantial thermal output, DGX units are equipped with heatsinks and fans designed to maintain optimal operating temperatures.

This framework makes DGX units suitable for computational tasks associated with artificial intelligence and machine learning models.[according to whom?]

Models

[edit]

Pascal - Volta

[edit]

DGX-1

[edit]

DGX-1 servers feature 8GPUs based on thePascal orVoltadaughter cards[1] with 128 GB of totalHBM2 memory, connected by anNVLinkmesh network.[2] The DGX-1 was announced on 6 April 2016.[3] All models are based on a dual socket configuration of Intel Xeon E5 CPUs, and are equipped with the following features.

  • 512 GB ofDDR4-2133
  • Dual 10 Gb networking
  • 4 x 1.92 TBSSDs
  • 3200W of combined power supply capability
  • 3U Rackmount Chassis

The product line is intended to bridge the gap between GPUs andAI accelerators using specific features for deep learning workloads.[4] The initial Pascal-based DGX-1 delivered 170teraflops ofhalf precision processing,[5] while the Volta-based upgrade increased this to 960teraflops.[6]

The DGX-1 was first available in only the Pascal-based configuration, with the first generation SXM socket. The later revision of the DGX-1 offered support for first generation Volta cards via the SXM-2 socket. Nvidia offered upgrade kits that allowed users with a Pascal-based DGX-1 to upgrade to a Volta-based DGX-1.[7][8]

  • The Pascal-based DGX-1 has two variants, one with a 16 coreIntel Xeon E5-2698 V3, and one with a 20 core E5-2698 V4. Pricing for the variant equipped with an E5-2698 V4 is unavailable, the Pascal-based DGX-1 with an E5-2698 V3 was priced at launch at $129,000[9]
  • The Volta-based DGX-1 is equipped with an E5-2698 V4 and was priced at launch at $149,000.[9]

DGX Station

[edit]

Designed as aturnkey deskside AI supercomputer, the DGX Station is atower computer that can function completely independently without typical datacenter infrastructure such as cooling, redundant power, or19 inch racks.

The DGX station was first available with the following specifications.[10]

  • Four Volta-based Tesla V100 accelerators, each with 16 GB ofHBM2 memory
  • 480 TFLOPS FP16
  • Single Intel Xeon E5-2698 v4[11]
  • 256 GB DDR4
  • 4x 1.92 TB SSDs
  • Dual 10 Gb Ethernet

The DGX station iswater-cooled to better manage the heat of almost 1500W of total system components, this allows it to keep a noise range under 35 dB under load.[12] This, among other features, made this system a compelling purchase for customers without the infrastructure to runrackmount DGX systems, which can be loud, output a lot of heat, and take up a large area. This was Nvidia's first venture into bringinghigh performance computing deskside, which has since remained a prominent marketing strategy for Nvidia.[13]

DGX-2

[edit]

The Nvidia DGX-2, the successor to the DGX-1, uses sixteen Volta-based V100 32 GB (second generation) cards in a single unit. It was announced on 27 March 2018.[14] The DGX-2 delivers 2 Petaflops with 512 GB of shared memory for tackling massive datasets and uses NVSwitch for high-bandwidth internal communication. DGX-2 has a total of 512 GB ofHBM2 memory, a total of 1.5 TB ofDDR4. Also present are eight 100 Gbit/sInfiniBand cards and 30.72 TB of SSD storage,[15] all enclosed within a massive 10U rackmount chassis and drawing up to 10 kW under maximum load.[16] The initial price for the DGX-2 was $399,000.[17]

The DGX-2 differs from other DGX models in that it contains two separate GPU daughterboards, each with eight GPUs. These boards are connected by an NVSwitch system that allows for full bandwidth communication across all GPUs in the system, without additional latency between boards.[16]

A higher performance variant of the DGX-2, the DGX-2H, was offered as well. The DGX-2H replaced the DGX-2's dual Intel Xeon Platinum 8168's with upgraded dual Intel Xeon Platinum 8174's. This upgrade does not increase core count per system, as both CPUs are 24 cores, nor does it enable any new functions of the system, but it does increase the base frequency of the CPUs from 2.7 GHz to 3.1 GHz.[18][19][20]

Ampere

[edit]

DGX A100 Server

[edit]

Announced and released on May 14, 2020. The DGX A100 was the 3rd generation of DGX server, including 8Ampere-based A100 accelerators.[21] Also included is 15 TB ofPCIe gen 4NVMe storage,[22] 1 TB of RAM, and eightMellanox-powered 200 GB/s HDR InfiniBand ConnectX-6NICs. The DGX A100 is in a much smaller enclosure than its predecessor, the DGX-2, taking up only 6 Rack units.[23]

The DGX A100 also moved to a 64 coreAMD EPYC 7742 CPU, the first DGX server to not be built with an Intel Xeon CPU. The initial price for the DGX A100 Server was $199,000.[21]

DGX Station A100

[edit]

As the successor to the original DGX Station, the DGX Station A100, aims to fill the same niche as the DGX station in being a quiet, efficient, turnkeycluster-in-a-box solution that can be purchased, leased, or rented by smaller companies or individuals who want to utilize machine learning. It follows many of the design choices of the original DGX station, such as the tower orientation, single socket CPUmainboard, a new refrigerant-based cooling system, and a reduced number of accelerators compared to the corresponding rackmount DGX A100 of the same generation.[13] The price for the DGX Station A100 320G is $149,000 and $99,000 for the 160G model, Nvidia also offers Station rental at ~US$9000 per month through partners in the US (rentacomputer.com) and Europe (iRent IT Systems) to help reduce the costs of implementing these systems at a small scale.[24][25]

The DGX Station A100 comes with two different configurations of the built in A100.

  • Four Ampere-based A100 accelerators, configured with 40 GB (HBM) or 80 GB (HBM2e) memory,
    thus giving a total of 160 GB or 320 GB resulting either in DGX Station A100 variants 160G or 320G.
  • 2.5 PFLOPS FP16
  • Single 64 CoreAMD EPYC 7742
  • 512 GBDDR4
  • 1 x 1.92 TBNVMe OS drive
  • 1 x 7.68 TB U.2 NVMe Drive
  • Dual port 10 Gb Ethernet
  • Single port 1 Gb BMC port

Hopper

[edit]

DGX H100 Server

[edit]

Announced March 22, 2022[26] and planned for release in Q3 2022,[27] The DGX H100 is the 4th generation of DGX servers, built with 8Hopper-based H100 accelerators, for a total of 32 PFLOPs of FP8 AI compute and 640 GB of HBM3 Memory, an upgrade over the DGX A100s 640GB HBM2 memory. This upgrade also increasesVRAM bandwidth to 3 TB/s.[28] The DGX H100 increases therackmount size to 8U to accommodate the 700W TDP of each H100 SXM card. The DGX H100 also has two 1.92 TB SSDs forOperating System storage, and 30.72 TB ofSolid state storage for application data.

One more notable addition is the presence of two NvidiaBluefield 3DPUs,[29] and the upgrade to 400 Gbit/s InfiniBand viaMellanox ConnectX-7NICs, double the bandwidth of the DGX A100. The DGX H100 uses new 'Cedar Fever' cards, each with four ConnectX-7 400 GB/s controllers, and two cards per system. This gives the DGX H100 3.2 Tbit/s of fabric bandwidth across Infiniband.[30]

The DGX H100 has two Xeon Platinum 8480C Scalable CPUs (CodenamedSapphire Rapids)[31] and 2 Terabytes ofSystem Memory.[32]

The DGX H100 was priced at £379,000 or ~US$482,000 at release.[33]

DGX GH200

[edit]

Announced May 2023, the DGX GH200 connects 32 Nvidia Hopper Superchips into a singular superchip, that consists totally of 256 H100 GPUs, 32 Grace Neoverse V2 72-core CPUs, 32 OSFT single-port ConnectX-7 VPI of with 400 Gbit/s InfiniBand and 16 dual-portBlueField-3 VPI with 200 Gbit/s ofMellanox[1][2] . Nvidia DGX GH200 is designed to handle terabyte-class models for massive recommender systems, generative AI, and graph analytics, offering 19.5 TB of shared memory with linear scalability for giant AI models.[34]

DGX Helios

[edit]

Announced May 2023, the DGX Helios supercomputer features 4 DGX GH200 systems. Each is interconnected with Nvidia Quantum-2 InfiniBand networking to supercharge data throughput for training large AI models. Helios includes 1,024 H100 GPUs.

Blackwell

[edit]

DGX GB200

[edit]
DGX B200/8 GPU
Nvidia DGX B200 8 way GPU Board (air cooled)
Nvidia GB200
Nvidia GB200 72 GPU liquid cooled rack system

Announced March 2024,[35] GB200 NVL72 connects 36 Grace Arm Neoverse V2 72-core CPUs and 72 B200 GPUs in a rack-scale design.[36] The GB200 NVL72 is a liquid-cooled, rack-scale solution that boasts a 72-GPU NVLink domain that acts as a single massive GPU.[37] Nvidia DGX GB200 offers 13.5 TB HBM3e of shared memory with linear scalability for giant AI models, less than its predecessor DGX GH200.

DGX SuperPod

[edit]

The DGX Superpod is a high performance turnkeysupercomputer system provided by Nvidia using DGX hardware.[38] It combines DGX compute nodes with fast storage and high bandwidthnetworking to provide a solution to high demand machine learning workloads. TheSelene supercomputer, at theArgonne National Laboratory, is one example of a DGX SuperPod-based system.

Selene, built from 280 DGX A100 nodes, ranked 5th on theTOP500 list for most powerful supercomputers at the time of its completion in June 2020,[39] and has continued to remain high in performance[citation needed]. The new Hopper-based SuperPod can scale to 32 DGX H100 nodes, for a total of 256 H100 GPUs and 64 x86 CPUs. This gives the complete SuperPod 20 TB of HBM3 memory, 70.4 TB/s of bisection bandwidth, and up to 1ExaFLOP ofFP8 AI compute.[28] These SuperPods can then be further joined to create larger supercomputers.

The Eos supercomputer, designed, built, and operated by Nvidia,[40][41][42] was constructed of 18 H100-based SuperPods, totaling 576 DGX H100 systems, 500 Quantum-2InfiniBand switches, and 360 NVLink Switches, that allow Eos to deliver 18 EFLOPs of FP8 compute, and 9 EFLOPs of FP16 compute, making Eos the 5th fastest AI supercomputer in the world, according to TOP500 (November 2023 edition).

As Nvidia does not produce any storage devices or systems, Nvidia SuperPods rely on partners to provide high performance storage. Current storage partners for Nvidia Superpods areDell EMC,DDN,HPE,IBM,NetApp, Pavilion Data, andVAST Data.[43]

DGX Spark

[edit]

In March 2025, Nvidia also announced the DGX Spark (previously DIGITS), a "desktop AI Supercomputer" based on Blackwell. These machines are targeted at AI researchers and programmers and have 128 GB of integrated RAM, making it possible to train or fine-tune fairly large models ("up to 200 billion parameters" with quantization). Several partner manufacturers also offer versions of the DGX Spark. It is available as of late 2025.[44][45]

Accelerators

[edit]

Comparison of accelerators used in DGX:[46][47][48]

ModelArchitectureSocketFP32
CUDA
cores
FP64 cores
(excl. tensor)
Mixed
INT32/FP32
cores
INT32
cores
Boost
clock
Memory
clock
Memory
bus width
Memory
bandwidth
VRAMSingle
precision
(FP32)
Double
precision
(FP64)
INT8
(non-tensor)
INT8
dense tensor
INT32FP4
dense tensor
FP16FP16
dense tensor
bfloat16
dense tensor
TensorFloat-32
(TF32)
dense tensor
FP64
dense tensor
Interconnect
(NVLink)
GPUL1 CacheL2 CacheTDPDie sizeTransistor
count
ProcessLaunched
P100PascalSXM/SXM235841792N/AN/A1480 MHz1.4 Gbit/s HBM24096-bit720 GB/sec16 GB HBM210.6 TFLOPS5.3 TFLOPSN/AN/AN/AN/A21.2 TFLOPSN/AN/AN/AN/A160 GB/secGP1001344 KB (24 KB × 56)4096 KB300 W610 mm215.3 BTSMC 16FF+Q2 2016
V100 16GBVoltaSXM251202560N/A51201530 MHz1.75 Gbit/s HBM24096-bit900 GB/sec16 GB HBM215.7 TFLOPS7.8 TFLOPS62 TOPSN/A15.7 TOPSN/A31.4 TFLOPS125 TFLOPSN/AN/AN/A300 GB/secGV10010240 KB (128 KB × 80)6144 KB300 W815 mm221.1 BTSMC 12FFNQ3 2017
V100 32GBVoltaSXM351202560N/A51201530 MHz1.75 Gbit/s HBM24096-bit900 GB/sec32 GB HBM215.7 TFLOPS7.8 TFLOPS62 TOPSN/A15.7 TOPSN/A31.4 TFLOPS125 TFLOPSN/AN/AN/A300 GB/secGV10010240 KB (128 KB × 80)6144 KB350 W815 mm221.1 BTSMC 12FFN
A100 40GBAmpereSXM4691234566912N/A1410 MHz2.4 Gbit/s HBM25120-bit1.52 TB/sec40 GB HBM219.5 TFLOPS9.7 TFLOPSN/A624 TOPS19.5 TOPSN/A78 TFLOPS312 TFLOPS312 TFLOPS156 TFLOPS19.5 TFLOPS600 GB/secGA10020736 KB (192 KB × 108)40960 KB400 W826 mm254.2 BTSMC N7Q1 2020
A100 80GBAmpereSXM4691234566912N/A1410 MHz3.2 Gbit/s HBM2e5120-bit1.52 TB/sec80 GB HBM2e19.5 TFLOPS9.7 TFLOPSN/A624 TOPS19.5 TOPSN/A78 TFLOPS312 TFLOPS312 TFLOPS156 TFLOPS19.5 TFLOPS600 GB/secGA10020736 KB (192 KB × 108)40960 KB400 W826 mm254.2 BTSMC N7
H100HopperSXM516896460816896N/A1980 MHz5.2 Gbit/s HBM35120-bit3.35 TB/sec80 GB HBM367 TFLOPS34 TFLOPSN/A1.98 POPSN/AN/AN/A990 TFLOPS990 TFLOPS495 TFLOPS67 TFLOPS900 GB/secGH10025344 KB (192 KB × 132)51200 KB700 W814 mm280 BTSMC 4NQ3 2022
H200HopperSXM516896460816896N/A1980 MHz6.3 Gbit/s HBM3e6144-bit4.8 TB/sec141 GB HBM3e67 TFLOPS34 TFLOPSN/A1.98 POPSN/AN/AN/A990 TFLOPS990 TFLOPS495 TFLOPS67 TFLOPS900 GB/secGH10025344 KB (192 KB × 132)51200 KB1000 W814 mm280 BTSMC 4NQ3 2023
B100BlackwellSXM6N/AN/AN/AN/AN/A8 Gbit/s HBM3e8192-bit8 TB/sec192 GB HBM3eN/AN/AN/A3.5 POPSN/A7 PFLOPSN/A1.98 PFLOPS1.98 PFLOPS989 TFLOPS30 TFLOPS1.8 TB/secGB100N/AN/A700 WN/A208 BTSMC 4NPQ4 2024
B200BlackwellSXM6N/AN/AN/AN/AN/A8 Gbit/s HBM3e8192-bit8 TB/sec192 GB HBM3eN/AN/AN/A4.5 POPSN/A9 PFLOPSN/A2.25 PFLOPS2.25 PFLOPS1.2 PFLOPS40 TFLOPS1.8 TB/secGB100N/AN/A1000 WN/A208 BTSMC 4NP

See also

[edit]

References

[edit]
  1. ^"nvidia dgx-1"(PDF). Retrieved15 November 2023.
  2. ^"inside pascal". 5 April 2016.Eight GPU hybrid cube mesh architecture with NVLink
  3. ^"NVIDIA Unveils the DGX-1 HPC Server: 8 Teslas, 3U, Q2 2016". Archived fromthe original on 6 April 2016.
  4. ^"deep learning supercomputer". 5 April 2016.
  5. ^"DGX-1 deep learning system"(PDF).NVIDIA DGX-1 Delivers 75X Faster Training...Note: Caffe benchmark with AlexNet, training 1.28M images with 90 epochs
  6. ^"DGX Server".DGX Server. Nvidia. Retrieved7 September 2017.
  7. ^Volta architecture whitepaper nvidia.com
  8. ^Use Guide nvidia.com
  9. ^abOh, Nate."NVIDIA Ships First Volta-based DGX Systems".www.anandtech.com. Archived fromthe original on 7 September 2017. Retrieved24 March 2022.
  10. ^"NVIDIA DGX Station Deep Learning System".www.compecta.com. Retrieved24 March 2022.
  11. ^"Intel® Xeon® Processor E5-2698 v4 (50M Cache, 2.20 GHz) - Product Specifications".Intel. Retrieved19 August 2023.
  12. ^Supercomputer datasheet nvidia.com
  13. ^ab"NVIDIA DGX Platform".NVIDIA. Retrieved15 November 2023.
  14. ^"Nvidia launches the DGX-2 with two petaFLOPS of power". 28 March 2018.
  15. ^"NVIDIA DGX -2 for Complex AI Challenges".NVIDIA. Retrieved24 March 2022.
  16. ^abCutress, Ian."NVIDIA's DGX-2: Sixteen Tesla V100s, 30 TB of NVMe, only $400K".www.anandtech.com. Archived fromthe original on 27 March 2018. Retrieved28 April 2022.
  17. ^"The NVIDIA DGX-2 is the world's first 2-petaflop single server supercomputer".www.hardwarezone.com.sg. 28 March 2018. Retrieved24 March 2022.
  18. ^DGX2 User Guide nvidia.com
  19. ^"Product Specifications".www.intel.com. Retrieved28 April 2022.
  20. ^"Product Specifications".www.intel.com. Retrieved28 April 2022.
  21. ^abSmith, Ryan (14 May 2020)."NVIDIA Ampere Unleashed: NVIDIA Announces New GPU Architecture, A100 GPU, and Accelerator". AnandTech. Archived fromthe original on 14 May 2020.
  22. ^Warren, Tom; Vincent, James (14 May 2020)."Nvidia's first Ampere GPU is designed for data centers and AI, not your PC". The Verge.
  23. ^"Boston Labs welcomes the DGX A100 to our remote testing portfolio!".www.boston.co.uk. Retrieved24 March 2022.
  24. ^Sharma, Mayank (13 April 2021)."Nvidia will let you rent its mini supercomputers".TechRadar. Retrieved31 March 2022.
  25. ^Walton, Jarred (12 April 2021)."Nvidia Refreshes Expensive, Powerful DGX Station 320G and DGX Superpod".Tom's Hardware. Retrieved28 April 2022.
  26. ^"NVIDIA Announces DGX H100 Systems – World's Most Advanced Enterprise AI Infrastructure".NVIDIA Newsroom. Retrieved24 March 2022.
  27. ^Albert (24 March 2022)."NVIDIA H100: Overview, Specs, & Release Date".www.seimaxim.com. Retrieved22 August 2022.
  28. ^abWalton, Jarred (22 March 2022)."Nvidia Reveals Hopper H100 GPU With 80 Billion Transistors".Tom's Hardware. Retrieved24 March 2022.
  29. ^"NVIDIA Announces DGX H100 Systems – World's Most Advanced Enterprise AI Infrastructure".NVIDIA Newsroom Newsroom. Retrieved19 April 2022.
  30. ^servethehome (14 April 2022)."NVIDIA Cedar Fever 1.6Tbps Modules Used in the DGX H100".ServeTheHome. Retrieved19 April 2022.
  31. ^"NVIDIA DGX H100 Datasheet".www.nvidia.com. Retrieved2 August 2023.
  32. ^"NVIDIA DGX H100".NVIDIA. Retrieved24 March 2022.
  33. ^Every NVIDIA DGX benchmarked & power efficiency & value compared, including the latest DGX H100., 8 December 2022, retrieved1 March 2023
  34. ^"NVIDIA DGX GH200".NVIDIA. Retrieved24 March 2022.
  35. ^"Nvidia CEO Jensen Huang announces new AI chips: 'We need bigger GPUs'".CNBC. 18 March 2024.
  36. ^"Supermicro NVIDIA GB200 NVL72 SuperCluster"(PDF). Supermicro. p. 4. Retrieved1 November 2025.
  37. ^"Nvidia Gb200 Nvl72".
  38. ^"NVIDIA SuperPOD Datasheet".NVIDIA. Retrieved15 November 2023.
  39. ^Trader, Tiffany (22 June 2020)."Nvidia Nabs #7 Spot on Top500 with Selene, Launches A100 PCIe Cards".HPC wire. Retrieved16 January 2025.
  40. ^Vincent, James (22 March 2022)."Nvidia reveals H100 GPU for AI and teases 'world's fastest AI supercomputer'".The Verge. Retrieved16 May 2022.
  41. ^Mellor, Chris (31 March 2022)."Nvidia Eos AI supercomputer will need a monster storage system".Blocks and Files. Retrieved21 May 2022.
  42. ^Moss, Sebastian (23 March 2022)."Nvidia announces Eos, "world's fastest AI supercomputer"".Data Center Dynamics. Retrieved21 May 2022.
  43. ^Mellor, Chris (31 March 2022)."Nvidia Eos AI supercomputer will need a monster storage system".Blocks and Files. Retrieved29 April 2022.
  44. ^"NVIDIA Announces DGX Spark and DGX Station Personal AI Computers".Nvidia press release. 18 March 2025. Retrieved28 June 2025.
  45. ^"NVIDIA DGX Spark 4 TB".DGX Spark in an online shop. 30 November 2025. Retrieved30 November 2025.
  46. ^Smith, Ryan (22 March 2022)."NVIDIA Hopper GPU Architecture and H100 Accelerator Announced: Working Smarter and Harder".AnandTech. Archived fromthe original on 23 September 2023.
  47. ^Smith, Ryan (14 May 2020)."NVIDIA Ampere Unleashed: NVIDIA Announces New GPU Architecture, A100 GPU, and Accelerator". AnandTech. Archived fromthe original on 29 July 2024.
  48. ^Garreffa, Anthony (17 September 2017)."NVIDIA Tesla V100 Tested: Near Unbelievable GPU Power".TweakTown.com. Retrieved30 December 2025.
Fixed pixel pipeline
Pre-GeForce
Vertex andpixel shaders
Unified shaders
Unified shaders &NUMA
Ray tracing &Tensor Cores
Software and technologies
Multimedia acceleration
Software
Technologies
GPU microarchitectures
Other products
GraphicsWorkstation cards
GPGPU software
Console components
Nvidia Shield
SoCs and embedded
CPUs
Computerchipsets
Company
Key people
Acquisitions
Retrieved from "https://en.wikipedia.org/w/index.php?title=Nvidia_DGX&oldid=1336112421"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp