计算机科学 ›› 2024, Vol. 51 ›› Issue (3): 39-47.doi: 10.11896/jsjkx.230700186

• 新计算模式下的信息安全防护 • 上一篇    下一篇

基于区块链的联邦蒸馏数据共享模型研究

刘炜1,2,3, 刘宇昭1,3, 唐琮轲1,3, 王媛媛5, 佘维1,3,4, 田钊1,3   

  1. 1 郑州大学网络空间安全学院 郑州450002
    2 河南省网络密码技术重点实验室(信息工程大学) 郑州450000
    3 郑州市区块链与数据智能重点实验室(郑州大学) 郑州450000
    4 嵩山实验室 郑州450000
    5 国网许昌供电公司 河南 许昌461000
  • 收稿日期:2023-07-24 修回日期:2023-11-30 出版日期:2024-03-15 发布日期:2024-03-13
  • 通讯作者: 田钊(tianzhao@zzu.edu.cn)
  • 作者简介:(wliu@zzu.edu.cn)
  • 基金资助:
    河南省高校科技创新人才支持计划(21HASTIT031);河南省网络密码技术重点实验室研究课题(LNCT2022-A04);河南省高等学校重点科研项目(24A520045);嵩山实验室预研项目(YYYY022022003)

Study on Blockchain Based Federated Distillation Data Sharing Model

LIU Wei1,2,3, LIU Yuzhao1,3, TANG Congke1,3, WANG Yuanyuan5, SHE Wei1,3,4, TIAN Zhao1,3   

  1. 1 School of Cyber Science and Engineering,Zhengzhou University,Zhengzhou 450002,China
    2 Henan Key Laboratory of Network Cryptography Technology(Information Engineering University),Zhengzhou 450000,China
    3 Zhengzhou Key Laboratory of Blockchain and Data Intelligence(Zhengzhou University),Zhengzhou 450000,China
    4 Songshan Laboratory,Zhengzhou 450000,China
    5 State Grid Henan Electric Power Company,Xuchang,Henan 461000,China
  • Received:2023-07-24 Revised:2023-11-30 Online:2024-03-15 Published:2024-03-13
  • About author:LIU Wei,born in 1981,Ph.D,associate professor,Ph.D supervisor,is a member of CCF(No.49811M).His main research interests include blockchain technology,privacy protection and smart healthcare.TIAN Zhao,born in 1985,Ph.D, asso-ciate professor,is a member of CCF(No.K7436M).His main research interests include blockchain technology,information security,and intelligent transport.
  • Supported by:
    Henan Province University Science and Technology Innovation Talent Support Plan(21HASTIT031),Research Project of Henan Provincial Key Laboratory of Network Cryptography Technology(LNCT2022-A04), Key Scientific Research Project of Colleges and Universities in Henan Province(24A520045) and Songshan Laboratory Pre-research Project(YYYY022022003)

摘要: 零散、孤立的海量数据形成“数据孤岛”使得数据无法交互和连接,如何在保护原始数据隐私的前提下安全有效地共享数据中的知识信息已成为热点研究问题。基于以上内容,提出了一种基于区块链的联邦蒸馏数据共享模型(BFDS)。区别于中心化架构,采用区块链联合多参与方组建教师网络,实现分布式协同工作;通过交换蒸馏输出的方式,传递数据中的知识信息,联合训练轻量化模型;提出了一种多权重节点可信评估算法,调用智能合约分配权重并生成可溯源全局软标签,降低因参与方质量差异而产生的负向影响。实验结果表明,BFDS模型能联合多参与方安全可信共享数据知识,协同蒸馏训练模型,降低了模型的部署成本;所提出的多权重节点评估算法能有效减小低质量节点的负向影响,提高了全局软标签的质量与安全性。

关键词: 区块链, 知识蒸馏, 数据共享, 智能合约

Abstract: The privacy of raw data makes it difficult to be directly shared among multiple participants.The issue of data security sharing and privacy-preserving has become a hot research topic.To solve this problem,this paper proposes a blockchain-based bederated distillation data sharing model(BFDS).It utilizes blockchain to form a collaborative teacher network with multiple participants.Through distilled output exchange,the knowledge from complex teacher networks is transferred and used to train lightweight models.A novel multi-weight node trust evaluation algorithm is proposed that uses smart contracts to generate traceable global soft labels.It can reduce the negative impact caused by quality differences among participants.Experimental results show that BFDS can collaborate with multiple parties to share data knowledge reliably,distill training models collaboratively,and reduce model deployment costs.The proposed algorithm can effectively reduce the negative impact of low-quality nodes and improve the quality and security of global soft labels.

Key words: Blockchain, Knowledge distillation, Data sharing, Smart contracts

中图分类号: 

  • TP391
[1]SAMUEL A L.Some studies in machine learning using the game of checkers [J]. IBM Journal of Research and Development,2000,44(1/2):206-226.
[2]MCMAHAN H B,MOORE E,RAMAGE D,et al.Federatedlearning of deep networks using model averaging[J].arXiv:1602.05629,2016,2.
[3]ZHOU X,XU M,WU Y,et al.Deep model poisoning attack on federated learning[J].Future Internet,2021,13(3):73.
[4]LIU W,TANG C K,MA J,et al.Application Research and Progress of Blockchain in Privacy Computing [J].Journal of Zhengzhou University(Natural Science Edition),2021,38(3):675-679.
[5]YUAN Y,WANG F Y.Development status and prospect ofblockchain technology[J].Acta Automatica Sinica,2016,42(4):481-494.
[6]MORA A,TENISON I,BELLAVISTA P,et al.Knowledge Distillation for Federated Learning:a Practical Guide[J].arXiv:2211.04742,2022.
[7]GEOFFREY H,ORIOL V,JEFF D.Distilling the knowledge in a neural network[J].arXiv:1503.02531,2015.
[8]WU C,WU F,LYU L,et al.Communication-efficient federated learning via knowledge distillation[J].Nature Communications,2022,13(1):2032.
[9]XU Z,HSU Y C,HUANG J.Training shallow and thin networks for acceleration via knowledge distillation with condi-tional adversarial networks[J].arXiv:1709.00513,2017.
[10]CHEN M,ZHANG L,MA T Y.Recommendation ApproachBased on Attentive Federated Distillation [J].Journal of Software,2021,32(12):3852-3868.
[11]SUN Y H,SHI Y H,LI M,et al.Personalized Federated Lear-ning Method Based on Collation Game and Knowledge Distillation[J].Journal of Electronics & Information Technology,2023,45(10):3702-3709.
[12]CHENG X M,DENG C H.Compression Algorithm of Face Re-cognition Model Based on Unlabeled Knowledge Distillation[J].Computer Science,2022,49(6):245-253.
[13]MO Z,GAO Z,ZHAO C,et al.FedDQ:A communication-efficient federated learning approach for Internet of Vehicles[J].Journal of Systems Architecture,2022,131:102690.
[14]YAO D,PAN W,DAI Y,et al.Local-global knowledge distil-lation in heterogeneous federated learning with non-iid data[J].arXiv:2107.00051,2021.
[15]LI N,ZHANG R,ZHU C,et al.A data sharing method for remote medical system based on federated distillation learning and consortium blockchain[J].Connection Science,2023,35(1):2186315.
[16]LI Y,ZHANG J,ZHU J,et al.HBMD-FL:Heterogeneous Fe-derated Learning Algorithm Based on Blockchain and Model Distillation[C]//Emerging Information Security and Applications:Third International Conference(EISA 2022).Cham:Springer Nature Switzerland,2023:145-159.
[17]MENG X F,LIU F,LI G,et al.Review of Knowledge Distillation in Convolutional Neural Network Compression[J].Journal of Frontiers of Computer Science and Technology,2021,15(10):1812-1829.
[18]LI H.Statistical learning methods[M].Beijing:Tsinghua University Press,2019.
[19]ZHU L G, LIU Z J,HAN S.Deep Leakage from Gradients[C]//Neural Information Processing Systems.2019.
[20]ZHOU C Y,CHEN D W,WANG S,et al.Research and Challenge of Distributed Deep Learning Privacy and Security Attack[J].Journal of Computer Research and Development,2021,58(5):927-943.
[21]HE Y Z,HU X B,HE J W,et al.Privacy and Security Issues in Machine Learning System:A Survey[J].Journal of Computer Research and Development,2019,56(10):2049-2070.
[22]QIU X Y,YE Z C,CUI X L,et al.Survey of communition overhead of federated learning[J].Journal of Computer Applications,2022,42(2):333-342.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!