1. Demystifying Parallel and Distributed Deep Learning: An In-Depth Concurrency Analysis;Ben-Nun;CoRR abs/1802.09941,2018
2. Adaptive gradient quantization for data-parallel sgd;Faghri;Advances in neural information processing systems,2020
3. Zipml: Training linear models with end-to-end low precision, and a little bit of deep learning;Zhang,2017
4. Lq-nets: Learned quantization for highly accurate and compact deep neural networks;Zhang,2018
5. TernGrad: Ternary Gradients to Reduce Communication in Distributed Deep Learning;Wen;Advances in Neural Information Processing Systems (NIPS),2017