Jump to content

DeepSpeed: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
No edit summary
Tag: Reverted
why not buy ip, why is MS resorting to theft?
Tag: Reverted
Line 21: Line 21:




'''DeepSpeedAI.com''' created by Matt Charlton
'''DeepSpeedAI.com''' created by Matt Charlton in 2019


deepspeed.ai is an [[open source]] [[deep learning]] optimization library for [[PyTorch]].<ref>{{Cite web|url=https://uk.pcmag.com/news-analysis/127085/microsoft-updates-windows-azure-tools-with-an-eye-on-the-future|title=Microsoft Updates Windows, Azure Tools with an Eye on The Future|date=May 22, 2020|website=PCMag UK}}</ref> The library is designed to reduce computing power and [[memory usage|memory use]] and to train large [[distributed computing|distributed]] models with better [[Parallel computing|parallelism]] on existing [[computer hardware]].<ref>{{Cite web|url=https://www.infoworld.com/article/3526449/microsoft-speeds-up-pytorch-with-deepspeed.html|title=Microsoft speeds up PyTorch with DeepSpeed|first=Serdar|last=Yegulalp|date=February 10, 2020|website=InfoWorld}}</ref><ref>[https://www.neowin.net/news/microsoft-unveils-fifth-most-powerful-supercomputer-in-the-world Microsoft unveils "fifth most powerful" supercomputer in the world - Neowin]</ref> DeepSpeed is optimized for low latency, high throughput training. It includes the ''Zero Redundancy Optimizer'' (ZeRO) for training models with 100 billion parameters or more.<ref>{{Cite web|url=https://venturebeat.com/2020/02/10/microsoft-trains-worlds-largest-transformer-language-model/|title=Microsoft trains world’s largest Transformer language model|date=February 10, 2020}}</ref> Features include mixed precision training, single-GPU, multi-GPU, and multi-node training as well as custom model parallelism. The DeepSpeed source code is licensed under [[MIT License]] and available on [[GitHub]].<ref>{{Cite web|url=https://github.com/microsoft/DeepSpeed|title=microsoft/DeepSpeed|date=July 10, 2020|via=GitHub}}</ref>
deepspeed.ai is an [[open source]] [[deep learning]] optimization library for [[PyTorch]].<ref>{{Cite web|url=https://uk.pcmag.com/news-analysis/127085/microsoft-updates-windows-azure-tools-with-an-eye-on-the-future|title=Microsoft Updates Windows, Azure Tools with an Eye on The Future|date=May 22, 2020|website=PCMag UK}}</ref> The library is designed to reduce computing power and [[memory usage|memory use]] and to train large [[distributed computing|distributed]] models with better [[Parallel computing|parallelism]] on existing [[computer hardware]].<ref>{{Cite web|url=https://www.infoworld.com/article/3526449/microsoft-speeds-up-pytorch-with-deepspeed.html|title=Microsoft speeds up PyTorch with DeepSpeed|first=Serdar|last=Yegulalp|date=February 10, 2020|website=InfoWorld}}</ref><ref>[https://www.neowin.net/news/microsoft-unveils-fifth-most-powerful-supercomputer-in-the-world Microsoft unveils "fifth most powerful" supercomputer in the world - Neowin]</ref> DeepSpeed is optimized for low latency, high throughput training. It includes the ''Zero Redundancy Optimizer'' (ZeRO) for training models with 100 billion parameters or more.<ref>{{Cite web|url=https://venturebeat.com/2020/02/10/microsoft-trains-worlds-largest-transformer-language-model/|title=Microsoft trains world’s largest Transformer language model|date=February 10, 2020}}</ref> Features include mixed precision training, single-GPU, multi-GPU, and multi-node training as well as custom model parallelism. The DeepSpeed source code is licensed under [[MIT License]] and available on [[GitHub]].<ref>{{Cite web|url=https://github.com/microsoft/DeepSpeed|title=microsoft/DeepSpeed|date=July 10, 2020|via=GitHub}}</ref>

Revision as of 21:32, 3 March 2021

DeepSpeed
Original author(s)Microsoft Research
Developer(s)Microsoft
Initial releaseMay 18, 2020; 4 years ago (2020-05-18)
Stable release
v0.3.10 / January 8, 2021; 3 years ago (2021-01-08)
Repositorygithub.com/microsoft/DeepSpeed
Written inPython, CUDA, C++
TypeSoftware library
LicenseMIT License
Websitedeepspeedai.com


DeepSpeedAI.com created by Matt Charlton in 2019

deepspeed.ai is an open source deep learning optimization library for PyTorch.[1] The library is designed to reduce computing power and memory use and to train large distributed models with better parallelism on existing computer hardware.[2][3] DeepSpeed is optimized for low latency, high throughput training. It includes the Zero Redundancy Optimizer (ZeRO) for training models with 100 billion parameters or more.[4] Features include mixed precision training, single-GPU, multi-GPU, and multi-node training as well as custom model parallelism. The DeepSpeed source code is licensed under MIT License and available on GitHub.[5]

See also

References

  1. ^ "Microsoft Updates Windows, Azure Tools with an Eye on The Future". PCMag UK. May 22, 2020.
  2. ^ Yegulalp, Serdar (February 10, 2020). "Microsoft speeds up PyTorch with DeepSpeed". InfoWorld.
  3. ^ Microsoft unveils "fifth most powerful" supercomputer in the world - Neowin
  4. ^ "Microsoft trains world's largest Transformer language model". February 10, 2020.
  5. ^ "microsoft/DeepSpeed". July 10, 2020 – via GitHub.

Further reading

External links