1. Deep gradient compression: Reducing the communication bandwidth for distributed training;Lin;ICLR,2018
2. Sparsified sgd with memory;Stich,2018
3. Qsgd: Communication-efficient sgd via gradient quantization and encoding;Alistarh;NeurIPS,2017
4. Fedpaq: A communication-efficient federated learning method with periodic averaging and quantization;Reisizadeh;AISTATS,2020
5. Rethinking gradient sparsification as total error minimization;Sahu;NeurIPS,2021