西安电子科技大学学报 ›› 2023, Vol. 50 ›› Issue (4): 121-131.doi: 10.19665/j.issn1001-2400.2023.04.012

• 网络空间安全专栏 • 上一篇    下一篇

隐私保护的拜占庭鲁棒联邦学习算法

李海洋1(),郭晶晶1(),刘玖樽1(),刘志全2,3()   

  1. 1.西安电子科技大学 网络与信息安全学院,陕西 西安 710071
    2.暨南大学 信息科学技术学院,广东 广州 510632
    3.数力聚(北京)科技有限公司,北京 100020
  • 收稿日期:2023-01-15 出版日期:2023-08-20 发布日期:2023-10-17
  • 通讯作者: 郭晶晶
  • 作者简介:李海洋(1996—),男,西安电子科技大学硕士研究生,E-mail:ocean5160@163.com;|刘玖樽(1997—),男,西安电子科技大学硕士研究生,E-mail:jzliu@stu.xidian.edu.cn;|刘志全(1989—),男,副研究员,E-mail:zqliu@jnu.edu.cn
  • 基金资助:
    陕西省自然科学基础研究计划(2022JQ-603);国家自然科学基金(62032025);国家自然科学基金(62272195);中央高校基本科研业务费专项资金(ZYTS23161);中央高校基本科研业务费专项资金(21622402);广东省网络与信息安全漏洞研究重点实验室项目(2020B1212060081);广州市科技计划项目(202201010421)

Privacy preserving byzantine robust federated learning algorithm

LI Haiyang1(),GUO Jingjing1(),LIU Jiuzun1(),LIU Zhiquan2,3()   

  1. 1. School of Cyber Engineering,Xidian University,Xi’an 710071,China
    2. College of Information Science and Technology,Jinan University,Guangzhou 510632,China
    3. Cyberdataforce (Beijing) Technology Ltd.,Beijing 100020,China
  • Received:2023-01-15 Online:2023-08-20 Published:2023-10-17
  • Contact: Jingjing GUO

摘要:

联邦学习是一种分布式机器学习范式,其中节点的原始训练集不出本地,它们通过共享模型更新来协作训练机器学习模型。当前联邦学习领域中的隐私保护和拜占庭攻击检测研究大都独立展开,现有的拜占庭攻击检测方法不可直接应用于隐私保护环境,不符合联邦学习的实际应用需求。针对上述问题,提出一种可在数据非独立同分布和隐私保护环境下拜占庭鲁棒的联邦学习算法。首先,以差分隐私技术为模型更新(本地模型梯度信息)提供隐私保护;然后,基于节点上传的历史模型更新对节点当前状态进行可信度评估;最后,根据评估结果进行全局模型聚合。仿真实验结果表明,在节点训练集非独立同分布、隐私保护和拜占庭节点比例为20%~80%的联邦学习环境中,所提算法进行拜占庭节点检测的漏检率和误检率均为0%。同时,随着节点数量的增加,拜占庭节点检测的时间开销呈线性增长的趋势。与现有的拜占庭节点检测算法相比,所提算法在节点数据非独立同分布及模型隐私保护情况下可得到更高精度的全局模型。

关键词: 联邦学习, 拜占庭攻击, 异常检测, 隐私保护技术, 差分隐私

Abstract:

Federated learning is a distributed machine learning paradigm,in which the original training sets of the nodes do not have to leave the local area and they collaborate to train machine learning models by sharing model updates.Most of the current privacy-preserving and Byzantine attack detection researches in the field of federated learning are carried out independently,and the existing Byzantine attack detection methods cannot be directly applied to the privacy-preserving environment,which does not meet the practical application requirements of federated learning.To address these problems,this paper proposes a federated learning algorithm for Byzantine robustness in a privacy-preserving environment with data non-independent and identically distributed.First,privacy protection is provided for model updates (local model gradient information) by differential privacy techniques; then the credibility is evaluated for the current state of nodes based on historical model updates uploaded by nodes; and finally,global model aggregation is performed based on the evaluation results.Simulation results show that in a federated learning environment with data non-independent and identically distributed,and with the privacy protection and Byzantine node ratio of 20%~80%,the proposed algorithm performs Byzantine node detection with both the miss detection rate and the false detection rate at 0%.Meanwhile,the time overhead of Byzantine node detection tends to linearly increase with the increase in the number of the nodes.Compared with the existing Byzantine node detection algorithms,the proposed algorithm can obtain a global model with a higher accuracy in the case of data being non-independent and identically distributed and model privacy protection.

Key words: federated learning, Byzantine attack, anomaly detection, privacy-preserving techniques, differential privacy

中图分类号: 

  • TP39
Baidu
map