1. SignSGD: Compressed optimisation for non-convex problems;bernstein;Proc Int Conf Mach Learn,2018
2. QSGD: Communication-efficient SGD via randomized quantization and encoding;alistarh;Proc Neural Inf Process Syst,2018
3. TernGrad: Ternary gradients to reduce communication in distributed deep learning;wen;Proc Neural Inf Process Syst,2017
4. Sparse Communication for Distributed Gradient Descent
5. Deep gradient compression: Reducing the communication bandwidth for distributed training;lin;Proc Int Conf Learn Represent,2018