Computer Science ›› 2026, Vol. 53 ›› Issue (4): 424-434.doi: 10.11896/jsjkx.250500116

• Information Security • Previous Articles     Next Articles

Lightweight Federated Continual Learning Method Based on Double Anti-forgetting Mechanism

WANG Pan, WANG Ji, ZHONG Zhengyi, BAO Weidong, ZHANG Yaohong   

  1. Laboratory for Big Data and Decision, National University of Defense Technology, Changsha 410000, China
  • Received:2025-05-26 Revised:2025-08-29 Online:2026-04-15 Published:2026-04-08
  • About author:WANG Pan,born in 2002,postgra-duate.His main research interests include federated learning and continual learning.
    WANG Ji,born in 1990,Ph.D,associate professor,master’s supervisor.His main research interests include deep learning and edge intelligence.
  • Supported by:
    National Natural Science Foundation of China(62002369).

Abstract: Federated learning(FL) enables knowledge sharing among different clients by uploading and aggregating client modelswithout sharing data.However,existing FL methods generally assume that client data is known and fixed.In reality,clients continuously receive tasks with new category data and update their models,which leads to a continuous decline in model performance on old tasks,known as catastrophic forgetting.To address this severe challenge,researchers have introduced continual learning(CL) into FL,giving rise to the research direction of federated continual learning(FCL).Nevertheless,as the number of tasks received by clients increases,existing FCL methods become less effective in alleviating catastrophic forgetting,especially for tasks that are relatively distant in time,where accuracy drops significantly.Moreover,the increasing degree of data heterogeneity further weakens model accuracy.To address this issue,this paper proposes a local-global anti-forgetting mechanism to mitigate the forgetting problem on distant tasks.Specifically,it introduces task-specific lightweight modules at the client level to effectively overcome catastrophic forgetting caused by data changes and model updates.At the server level,it generates and filters category-balanced pseudo-images through model inversion to alleviate the decline in model performance due to data distribution diffe-rences.Through a series of experiments conducted on CIFAR-10,CIFAR-100,and TinyImageNet datasets,the results strongly demonstrate the superiority of the proposed mechanism.Compared with existing methods,it shows significant advantages in improving model performance and alleviating catastrophic forgetting.

Key words: Federated continual learning, Catastrophic forgetting, Data heterogeneity, Lightweight modules, Local-global anti forgetting

CLC Number: 

  • TP301
[1]DE CAPITANI DI VIMERCATI S,FORESTI S,LIVRAGA G,et al.Data privacy:Definitions and techniques[J].International Journal of Uncertainty,Fuzziness and Knowledge-Based Systems,2012,20(6):793-817.
[2]PARDAU S L.The california consumer privacy act:Towards a European-style privacy regime in the united states[J].Journal of Technology Law & Policy,2018,23:68.
[3]VOIGT P,VON DEM BUSSCHE A.The EU general data protection regulation(GDPR)[M].Cham:Springer,2017.
[4]CHI J L,FENG D G,ZHANG M.,et al.Research Progress on Privacy-Preserving Ciphertext Retrieval Technology[J].Journal of Electronics and Information,2024,46(5):1-24.
[5]FIENBERG S E,SLAVKOVIĆ A B.Data privacy and confidentiality[M]//International Encyclopedia of Statistical Science.Berlin:Springer,2025:615-619.
[6]OMID P,SOREN F.The Digital Double:Data Privacy,Security,and Consent in AI Implants West[J].Journal of Dental Sciences,2025,2(1):108.
[7]BALOGUN A Y.Strengthening compliance with data privacyregulations in US healthcare cybersecurity[J].Asian Journal of Research in Computer Science,2025,18(1):154-173.
[8]MCMAHAN B,MOORE E,RAMAGE D,et al.Communica-tion efficient learning of deep networks from decentralized data[C]//Artificial Intelligence and Statistics.PMLR,2017:1273-1282.
[9]YANG Q.AI and Data Privacy Protection:Solution of Federated Learning[J].Information Security Research,2019,5(11):961.
[10]DEMBANI R,KARVELAS I,AKBAR N A,et al.Agricultural data privacy and federated learning:A review of challenges and opportunities[J].Computers and Electronics in Agriculture,2025,232:110048.
[11]KEMKER R,MCCLURE M,ABITINO A,et al.Measuring catastrophic forgetting in neural networks[C]//Proceedings of the AAAI Conference on Artificial Intelligence.2018.
[12]LI Y C,WANG H Z,XU W C,et al.Unleashing the power of continual learning on non-centralized devices:A survey[J].IEEE Communications Surveys & Tutorials,2025,28:1059-1098.
[13]QU H,RAHMANI H,XU L,et al.Recent advances of continual learning in computer vision:An overview[J].IET Computer Vision,2025,19(1):e70013.
[14]YANG X,YU H,GAO X,et al.Federated continual learning via knowledge fusion:A survey[J].IEEE Transactions on Know-ledge and Data Engineering,2024,36(8):3832-3850.
[15]YOON J,JEONG W,LEE G,et al.Federated continual learning with weighted inter-client transfer[C]//International Confe-rence on Machine Learning.PMLR,2021:12073-12086.
[16]ZHANG J,CHEN C,ZHUANG W,et al.Target:Federatedclass-continual learning via exemplar-free distillation[C]//Proceedings of the IEEE/CVF International Conference on Computer Vision.2023:4782-4793.
[17]FANG Z X,FU X D,DING J M,et al.Federated Class-incremental Learning for Unbalanced Data[J].Journal of Chinese Computer Systems,2025,46(9):2121-2129.
[18]TONG G,LI G,WU J,et al.GradMFL:Gradient memory-based federated learning for hierarchical knowledge transferring over non-iid data[C]//International Conference on Algorithms and Architectures for Parallel Processing.Cham:Springer,2021:612-626.
[19]ZHANG Z,ZHANG Y,GUO D,et al.Communication-efficient federated continual learning for distributed learning system with Non-IID data[J].Science China Information Sciences,2023,66(2):122102.
[20]CICIRELLO V,HU K,LU M,et al.A Federated Incremental Learning Algorithm Based on Dual Attention Mechanism[J].Applied Sciences,2022,12(19),10025.
[21]USMANOVA A,PORTET F,LALANDA P,et al.A distil-lation-based approach integrating continual learning and federated learning for pervasive services[J].arXiv:2109.04197,2021.
[22]BABAKNIYA S,FABIAN Z,HE C,et al.A data-free approach to mitigate catastrophic forgetting in federated class incremental learning for vision tasks[J].Advances in Neural Information Processing Systems,2023,36:66408-66425.
[23]GUBEROVIĆ E,ALEXOPOULOS C,BOSNIĆ I,et al.Frame-work for federated learning open models in e-government applications[J].Interdisciplinary Description of Complex Systems:INDECS,2022,20(2):162-178.
[24]FATHIMAA S,BASHA S M,AHMED S T,et al.Empowering consumer healthcare through sensor-rich devices using federated learning for secure resource recommendation[J].IEEE Transactions on Consumer Electronics,2025,71(1):1563-1570.
[25]KONEČNÝ J,MCMAHAN H B,YU F X,et al.Federated learning:Strategies for improving communication efficiency[J].arXiv:1610.05492,2016.
[26]HSIEH K,HARLAP A,VIJAYKUMAR N,et al.Gaia:{Geo-Distributed} machine learning approaching {LAN} speeds[C]//14th USENIX Symposium on Networked Systems Design and Implementation(NSDI 17).2017:629-647.
[27]LUPI NG W,WEI W,BO L I.CMFL:Mitigating communication overhead for federated learning[C]//2019 IEEE 39th International Conference on Distributed Computing Systems(ICDCS).IEEE,2019:954-964.
[28]BONAWITZ K,EICHNER H,GRIESKAMP W,et al.Towards federated learning at scale:System design[J].Proceedings of Machine Learning and Systems,2019,1:374-388.
[29]LIU Y,JAMES J Q,KANG J,et al.Privacy-preserving traffic flow prediction:A federated learning approach[J].IEEE Internet of Things Journal,2020,7(8):7751-7763.
[30]TANG L T,WANG D,ZHANG L F,et al.A Federated Lear-ning Scheme Based on Secure Multi-Party Computation and Differential Privacy[J].Computer Science,2022,49(9):297-305.
[31]SHEN W,HUANG W,WAN G,et al.Label-free backdoor attacks in vertical federated learning[C]//Proceedings of the AAAI Conference on Artificial Intelligence.2025:20389-20397.
[32]MAI Z,LI R,JEONG J,et al.Online continual learning in image classification:An empirical survey[J].Neurocomputing,2022,469:28-51.
[33]MALLYA A,LAZEBNIK S.Packnet:Adding multiple tasks to a single network by iterative pruning[C]//Proceedings of the IEEE Conference on Computer Visionand Pattern Recognition.2018:7765-7773.
[34]ALJUNDI R,CHAKRAVARTY P,TUYTELAARS T.Expert gate:Lifelong learning with a network of experts[C]//Procee-dings of the IEEE Conference on Computer Vision and Pattern Recognition.2017:3366-3375.
[35]ROLNICK D,AHUJA A,SCHWARZ J,et al.Experience replay for continual learning[C]//Proceedings of the 33rd Internatio-nal Conference on Neural Information Processing Systems.2019:350-360.
[36]SHIN H,LEE J K,KIM J,et al.Continual learning with deep generative replay[J].arXiv:1705.08690,2017.
[37]LI Z,HOIEM D.Learning without forgetting[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2017,40(12):2935-2947.
[38]KIRKPATRICK J,PASCANU R,RABINOWITZ N,et al.Overcoming catastrophic forgetting in neural networks[J].Proceedings of the National Academy of Sciences,2017,114(13):3521-3526.
[39]DONG J,WANG L,FANG Z,et al.Federated class-incremental learning[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.2022:10164-10173.
[40]ALEX KRIZHEVSKY.Cifar-10 and cifar-100 datasets[EB/OL].http://www.cs.toronto.edu/kriz/cifar.html.
[41]LE Y,YANG X.Tiny imagenet visual recognition challenge[EB/OL].http://vision.stanford.edu/teaching/cs231n/reports/2015/pdfs/yle_project.pdf.
[42]LI T,SAHU A K,ZAHEER M,et al.Federated optimization in heterogeneous networks[J].Proceedings of Machine Learning and Systems,2020,2:429-450.
[43]YU H,YANG X,ZHANG L,et al.Handling spatial-temporal data heterogeneity for federated continual learning via tail anchor[C]//Proceedings of the Computer Vision and Pattern Re-cognition Conference.2025:4874-4883.
[1] LI Linhao, XU Yanan, DONG Yongfeng, WANG Zhen. Social Learning Based on Multi-expert Collaboration and Information Interaction [J]. Computer Science, 2026, 53(3): 197-206.
[2] WU Jiagao, YI Jing, ZHOU Zehui, LIU Linfeng. Personalized Federated Learning Framework for Long-tailed Heterogeneous Data [J]. Computer Science, 2025, 52(9): 232-240.
[3] WANG Chundong, ZHANG Qinghua, FU Haoran. Federated Learning Privacy Protection Method Combining Dataset Distillation [J]. Computer Science, 2025, 52(6A): 240500132-7.
[4] WANG Dongzhi, LIU Yan, GUO Bin, YU Zhiwen. Edge-side Federated Continuous Learning Method Based on Brain-like Spiking Neural Networks [J]. Computer Science, 2025, 52(3): 326-337.
[5] XIE Jiachen, LIU Bo, LIN Weiwei , ZHENG Jianwen. Survey of Federated Incremental Learning [J]. Computer Science, 2025, 52(3): 377-384.
[6] ZENG Peiyi. Classification of Multiscale Steel Microstructure Images Based on Incremental Learning [J]. Computer Science, 2024, 51(6A): 230500180-8.
[7] XU Yicheng, DAI Chaofan, MA Wubin, WU Yahui, ZHOU Haohao, LU Chenyang. Particle Swarm Optimization-based Federated Learning Method for Heterogeneous Data [J]. Computer Science, 2024, 51(6): 391-398.
[8] ZHANG Yu, CAO Xiqing, NIU Saisai, XU Xinlei, ZHANG Qian, WANG Zhe. Incremental Class Learning Approach Based on Prototype Replay and Dynamic Update [J]. Computer Science, 2023, 50(11A): 230300012-7.
[9] LIU Dong-mei, XU Yang, WU Ze-bin, LIU Qian, SONG Bin, WEI Zhi-hui. Incremental Object Detection Method Based on Border Distance Measurement [J]. Computer Science, 2022, 49(8): 136-142.
[10] ZHOU Hai-yu, ZHANG Dao-qiang. Multi-site Hyper-graph Convolutional Neural Networks and Application [J]. Computer Science, 2022, 49(3): 129-133.
[11] XIE Yu, YANG Rui-ling, LIU Gong-xu, LI De-yu, WANG Wen-jian. Human Skeleton Action Recognition Algorithm Based on Dynamic Topological Graph [J]. Computer Science, 2022, 49(2): 62-68.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!