Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

DeepSpeed

From Wikipedia, the free encyclopedia
Microsoft open source library
DeepSpeed
Original author(s)Microsoft Research
Developer(s)Microsoft
Initial releaseMay 18, 2020; 4 years ago (2020-05-18)
Stable release
v0.16.2 / December 18, 2024; 2 months ago (2024-12-18)
Repositorygithub.com/microsoft/DeepSpeed
Written inPython,CUDA,C++
TypeSoftware library
LicenseApache License 2.0
Websitedeepspeed.ai

DeepSpeed is anopen sourcedeep learning optimization library forPyTorch.[1]

Library

[edit]

The library is designed to reduce computing power andmemory use and to train largedistributed models with betterparallelism on existingcomputer hardware.[2][3] DeepSpeed is optimized for low latency, high throughput training. It includes the Zero Redundancy Optimizer (ZeRO) for training models with 1 trillion or more parameters.[4] Features include mixed precision training, single-GPU, multi-GPU, and multi-node training as well as custom model parallelism. The DeepSpeed source code is licensed underMIT License and available onGitHub.[5]

The team claimed to achieve up to a 6.2x throughput improvement, 2.8x faster convergence, and 4.6x less communication.[6]

See also

[edit]

References

[edit]
  1. ^"Microsoft Updates Windows, Azure Tools with an Eye on The Future".PCMag UK. May 22, 2020.
  2. ^Yegulalp, Serdar (February 10, 2020)."Microsoft speeds up PyTorch with DeepSpeed".InfoWorld.
  3. ^"Microsoft unveils "fifth most powerful" supercomputer in the world".Neowin. 18 June 2023.
  4. ^"Microsoft trains world's largest Transformer language model". February 10, 2020.
  5. ^"microsoft/DeepSpeed". July 10, 2020 – via GitHub.
  6. ^"DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression".Microsoft Research. 2021-05-24. Retrieved2021-06-19.

Further reading

[edit]
  • Rajbhandari, Samyam; Rasley, Jeff; Ruwase, Olatunji; He, Yuxiong (2019). "ZeRO: Memory Optimization Towards Training A Trillion Parameter Models".arXiv:1910.02054 [cs.LG].

External links

[edit]
Open source
Proprietary
Overview
Software
Applications
Video games
Programming
languages
Frameworks,
development tools
Operating systems
Other
Licenses
Forges
Related
Main
projects
Languages, compilers
Distributedgrid computing
Internet,networking
Other projects
Operating systems
APIs
Launched as products
MSR Labs
applied
research
Live Labs
Current
Discontinued
FUSE Labs
Other labs
Retrieved from "https://en.wikipedia.org/w/index.php?title=DeepSpeed&oldid=1279611738"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp