Abstract
Federated incremental learning best suits the changing needs of common Federal Learning (FL) tasks. In this area, the large sample client dramatically influences the final model training results, and the unbalanced features of the client are challenging to capture. In this paper, a federated incremental learning framework is designed; firstly, part of the data is preprocessed to obtain the initial global model. Secondly, to help the global model to get the importance of the features of the whole sample of each client, and enhance the performance of the global model to capture the critical information of the feature, channel attention neural network model is designed on the client side, and a federated aggregation algorithm based on the feature attention mechanism is designed on the server side. Experiments on standard datasets CIFAR10 and CIFAR100 show that the proposed algorithm accuracy has good performance on the premise of realizing incremental learning.
Funder
National Natural Science Foundation of China
Subject
Fluid Flow and Transfer Processes,Computer Science Applications,Process Chemistry and Technology,General Engineering,Instrumentation,General Materials Science
Reference51 articles.
1. Communication-efficient learning of deep networks from decentralized data;McMahan;arXiv,2017
2. Federated Learning of Deep Networks using Model Averaging;McMahan;Electr. Power Syst. Res.,2017
3. Light GBM: A Highly Efficient Gradient Boosting Decision Tree;Qi;Adv. Neural Inf. Process. Syst.,2017
4. EGBMMDA: Extreme Gradient Boosting Machine for MiRNA-Disease Association prediction
5. Support Vector Machine;Saunders;Comput. Sci.,2002
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Federated Continual Learning via Knowledge Fusion: A Survey;IEEE Transactions on Knowledge and Data Engineering;2024-08
2. A Round-Based Network Attack Detection Model Using Auto-encoder In IoT-Edge Computing;2023 7th International Conference on Internet of Things and Applications (IoT);2023-10-25