Abstract
AbstractFederated learning has gained significant attention due to its groundbreaking ability to enable distributed learning while maintaining privacy constraints. However, as a consequence of data heterogeneity among decentralized devices, it inherently experiences significant learning degradation and slow convergence speed. Therefore, it is natural to employ the concept of clustering homogeneous clients into the same group, allowing only the model weights within each group to be aggregated. While most existing clustered federated learning methods employ either model gradients or inference outputs as metrics for client partitioning to group similar devices together, heterogeneity may still exist within each cluster. Moreover, there is a scarcity of research exploring the underlying reasons for determining the appropriate timing for clustering, resulting in the common practice of assigning each client to its own individual cluster, particularly in the context of highly non-independent and identically distributed (Non-IID) data. In this paper, we introduce a two-stage decoupling federated learning algorithm with adaptive personalization layers named FedTSDP, where client clustering is performed twice according to inference outputs and model weights, respectively. Hopkins amended sampling is adopted to determine the appropriate timing for clustering and the sampling weight of public unlabeled data. In addition, a simple yet effective approach is developed to adaptively adjust the personalization layers based on varying degrees of data skew. Experimental results show that our proposed method has reliable performance on both IID and non-IID scenarios.
Funder
National Natural Science Foundation of China
Publisher
Springer Science and Business Media LLC
Reference47 articles.
1. Al-Abiad MS, Obeed M, Hossain MJ, Chaaban A (2023) Decentralized aggregation for energy-efficient federated learning via D2D communications. IEEE Trans Commun 71(6):3333–3351. https://doi.org/10.1109/TCOMM.2023.3253718
2. Arivazhagan MG, Aggarwal V, Singh AK, Choudhary S (2019) Federated learning with personalization layers. arXiv preprint arXiv:1912.00818
3. Banerjee A, Dave RN (2004) Validating clusters using the Hopkins statistic. In: 2004 IEEE international conference on fuzzy systems (IEEE Cat. No. 04CH37542), vol 1. IEEE, pp 149–153 (2004)
4. Briggs C, Fan Z, Andras P (2020) Federated learning with hierarchical clustering of local updates to improve training on non-IID data. In: 2020 International joint conference on neural networks (IJCNN). IEEE, pp 1–9
5. Cho YJ, Wang J, Chirvolu T, Joshi G (2023) Communication-efficient and model-heterogeneous personalized federated learning via clustered knowledge transfer. IEEE J Sel Top Signal Process 17(1):234–247