Large-scale models are revolutionizing deep learning and AI research, driving major improvements in language understanding, generating creative texts, multi-lingual translation and many more. But despite their remarkable capabilities, the models’ large size creates latency and cost constraints that hinder the deployment of applications on top of them. In particular, increased inference time and memory consumption […]
如何评价微软开源的分布式训练框架deepspeed? - 菩提树的回答- 知乎
Latest News - DeepSpeed
This AI newsletter is all you need #6, by Towards AI Editorial Team
GitHub - microsoft/DeepSpeed: DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
deepspeed - Python Package Health Analysis
Compressing neural nets — The Dan MacKinlay stable of variably-well-consider'd enterprises
Optimization approaches for Transformers [Part 2]
ChatGPT只是前菜,2023要来更大的! - 墨天轮
PDF] DeepSpeed- Inference: Enabling Efficient Inference of Transformer Models at Unprecedented Scale
Amanuel Alambo (@amanuel_alambo) / X
ChatGPT只是前菜,2023要来更大的! - 墨天轮
ZeRO-2 & DeepSpeed: Shattering barriers of deep learning speed & scale - Microsoft Research
This AI newsletter is all you need #6 – Towards AI