1. [1]H. Wang, H. Tian, J. Chen, X. Wan, J. Xia, G. Zeng, W. Bai, J. Jiang, Y. Wang, and K. Chen. Towards Domain-specific Network Transport for Distributed DNN Training. In Proc. USENIX NSDI, 2024.
2. Scaling Distributed Machine Learning with the Parameter Server
3. [3]Z. Jia, M. Zaharia, and A. Aiken. Beyond Data and Model Parallelism for Deep Neural Networks. In Proc. MLSys, 2019.
4. PipeDream
5. [5]Y. Huang, Y. Cheng, A. Bapna, et al. Gpipe: Efficient Training of Giant Neural Networks Using Pipeline Parallelism. In Proc. NIPS, 2019.