Senior Research Scientist at Google DeepMind Japan office; ex-Microsoft DeepSpeed member; CMU Ph.D.
- Tokyo, Japan
- https://conglongli.github.io/
- @conglongli
Stars
Latency and Memory Analysis of Transformer Models for Training and Inference
deepspeedai / Megatron-DeepSpeed
Forked from NVIDIA/Megatron-LMOngoing research training transformer language models at scale, including: BERT & GPT-2
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Example models using DeepSpeed
Source code for SIGMOD 2020 paper "Improving Approximate Nearest Neighbor Search through Learned Adaptive Early Termination"