Jump to content
Wikipedia The Free Encyclopedia

DeepSpeed

From Wikipedia, the free encyclopedia
Microsoft open source library
DeepSpeed
Original author(s) Microsoft Research
Developer(s) Microsoft
Initial releaseMay 18, 2020; 4 years ago (2020年05月18日)
Stable release
v0.16.2 / December 18, 2024; 2 months ago (2024年12月18日)
Repository github.com/microsoft/DeepSpeed
Written inPython, CUDA, C++
Type Software library
License Apache License 2.0
Websitedeepspeed.ai

DeepSpeed is an open source deep learning optimization library for PyTorch.[1]

Library

[edit ]

The library is designed to reduce computing power and memory use and to train large distributed models with better parallelism on existing computer hardware.[2] [3] DeepSpeed is optimized for low latency, high throughput training. It includes the Zero Redundancy Optimizer (ZeRO) for training models with 1 trillion or more parameters.[4] Features include mixed precision training, single-GPU, multi-GPU, and multi-node training as well as custom model parallelism. The DeepSpeed source code is licensed under MIT License and available on GitHub.[5]

The team claimed to achieve up to a 6.2x throughput improvement, 2.8x faster convergence, and 4.6x less communication.[6]

See also

[edit ]

References

[edit ]

Further reading

[edit ]
  • Rajbhandari, Samyam; Rasley, Jeff; Ruwase, Olatunji; He, Yuxiong (2019). "ZeRO: Memory Optimization Towards Training A Trillion Parameter Models". arXiv:1910.02054 [cs.LG].
[edit ]
Overview
Software
Applications
Video games
Programming
languages
Frameworks,
development tools
Operating systems
Other
Licenses
Forges
Related
Main
projects
Languages, compilers
Distributedgrid computing
Internet, networking
Other projects
Operating systems
APIs
Launched as products
MSR Labs
applied
research
Live Labs
Current
Discontinued
FUSE Labs
Other labs

AltStyle によって変換されたページ (->オリジナル) /