计算机科学 ›› 2023, Vol. 50 ›› Issue (11): 356-363.doi: 10.11896/jsjkx.221200005
许文韬, 王斌君
XU Wentao, WANG Binjun
摘要: 联邦学习解决了用户隐私与数据共享相悖之大数据困局,体现了“数据可用不可见”的理念。然而,联邦模型在训练过程中存在后门攻击的风险。攻击者通过本地训练一个包含后门任务的攻击模型,并将模型参数放大一定比例,从而实现将后门植入联邦模型中。针对横向联邦学习模型所面临的后门威胁,从博弈的视角,提出一种基于随机断层与梯度剪裁相结合的后门防御策略和技术方案:中心服务器在收到参与方提交的梯度信息后,随机确定每个参与方的神经网络层,然后将各参与方的梯度贡献分层聚合,并使用梯度阈值对梯度参数进行裁剪。梯度剪裁和随机断层可削弱个别参与方异常数据的影响力,使联邦模型在学习后门特征时陷入平缓期,长时间无法学习到后门特征,同时不影响正常任务的学习。如果中心服务器在平缓期内结束联邦学习,即可实现对后门攻击的防御。实验结果表明,该方法可以有效地防御联邦学习中潜在的后门威胁,同时保证了模型的准确性。因此,该方法可以应用于横向联邦学习场景中,为联邦学习的安全保驾护航。
中图分类号:
[1]MCMAHAN B,MOORE E,RAMAGE D,et al.Communication-efficient learning of deep networks from decentralized data[C]//Artificial Intelligence and Statistics.Florida:PMLR,2017:1273-1282. [2]XU J,GLICKSBERG B S,SU C,et al.Federated learning for healthcare informatics[J].Journal of Healthcare Informatics Research,2021,5(1):1-19. [3]LIN B Y,HE C,ZENG Z,et al.Fednlp:Benchmarking federated learning methods for natural language processing tasks[C]//Findings of the Association for Computational Linguistics:NAACL 2022.Stroudsburg:ACL,2022:157-175. [4]BYRD D,POLYCHRONIADOU A.Differentially private secure multi-party computation for federated learning in financial applications[C]//Proceedings of the First ACM International Conference on AI in Finance.New York:ACM,2020:1-9. [5]KAIROUZ P,MCMAHAN H B,AVENT B,et al.Advancesand open problems in federated learning[J].Foundations and Trends in Machine Learning,2021,14(1/2):1-210. [6]TOLPEGIN V,TRUEX S,GURSOY M E,et al.Data poisoning attacks against federated learning systems[C]//European Symposium on Research in Computer Security.New York:Springer,2020:480-501. [7]WANG H,SREENIVASAN K,RAJPUT S,et al.Attack of the tails:Yes,you really can backdoor federated learning[J].Advances in Neural Information Processing Systems,2020,33:16070-16084. [8]GONG X,CHEN Y,HUANG H,et al.Coordinated Backdoor Attacks against Federated Learning with Model-Dependent Triggers[J].IEEE Network,2022,36(1):84-90. [9]BONAWITZ K,IVANOV V,KREUTER B,et al.Practical secure aggregation for privacy-preserving machine learning[C]//Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security.New York:ACM,2017:1175-1191. [10]SUN Z,KAIROUZ P,SURESH A T,et al.Can you really backdoor federated learning?[J].arXiv:1911.07963,2019. [11]GAO J,ZHANG B,GUO X,et al.Secure Partial Aggregation:Making Federated Learning More Robust for Industry 4.0 Applications[J].IEEE Transactions on Industrial Informatics,2022,18(9):6340-6348. [12]LI S H,ZHENG H B,CHEN J Y,et al.Neural Path Poisoning Attack Method for Federated Learning[J].Journal of Chinese Computer Systems,2023,44(7):1578-1585. [13]BAGDASARYAN E,VEIT A,HUA Y,et al.How to backdoor federated learning[C]//International Conference on Artificial Intelligence and Statistics.New York:PMLR,2020:2938-2948. [14]LIU Y,MA S,AAFER Y,et al.Trojaning attack on neural networks[C]//25th Annual Network and Distributed System Security Symposium.California:The Internet Society,2018:1-11. [15]ZHANG J,HE T,SRA S,et al.Why gradient clipping accele-rates training:A theoretical justification for adaptivity[J].ar-Xiv:1905.11881,2019. [16]CALDAS S,DUDDU S M K,WU P,et al.Leaf:A benchmark for federated settings[J].arXiv:1812.01097,2018. [17]LI Q,DIAO Y,CHEN Q,et al.Federated learning on non-iid data silos:An experimental study[C]//2022 IEEE 38th International Conference on Data Engineering(ICDE).New York:IEEE,2022:965-978. [18]ZHU L,HAN S.Deep leakage from gradients[C]//Advances in Neural Information Processing Systems 32:Annual Conference on Neural Information Processing Systems.New York:Curran Associates Inc,2019:14747-14756. |
|