1. Petuum: A framework for iterative-convergent distributed Ml;dai;arXiv 1312 7651,2013
2. More effective distributed Ml via a stale synchronous parallel parameter server;ho;Proc 27th Annu Conf Neural Inf Process Syst,2013
3. On variance reduction in stochastic gradient descent and its asynchronous variants;reddi;Proc Annu Conf Neural Inf Process Syst (NIPS),2015
4. Large scale distributed deep networks;dean;Proc 26th Annu Conf Neural Inf Process Syst,2012
5. Deep gradient compression: Reducing the communication bandwidth for distributed training;lin;arXiv 1712 01887,2017