1. Federated learning: Strategies for improving communication efficiency;Konečný,2016
2. Local SGD converges fast and communicates little;Stich
3. Federated accelerated stochastic gradient descent;Yuan;Advances in Neural Information Processing Systems,2020
4. Proxskip: Yes! local gradient steps provably lead to communication acceleration! finally!;Mishchenko
5. Cooperative SGD: A unified framework for the design and analysis of communication-efficient SGD algorithms;Wang;Journal of Machine Learning Research,2021