正确性
计算机科学
可验证秘密共享
稳健性(进化)
堆
数学证明
联合学习
分布式计算
人工智能
理论计算机科学
算法
生物化学
化学
几何学
数学
集合(抽象数据类型)
基因
程序设计语言
作者
Xiangyun Tang,Meng Shen,Qi Li,Liehuang Zhu,Tengfei Xue,Qiang Qu
标识
DOI:10.1109/tdsc.2023.3239007
摘要
Federated learning (FL) protects training data in clients by collaboratively training local machine learning models of clients for a global model, instead of directly feeding the training data to the server. However, existing studies show that FL is vulnerable to various attacks, resulting in training data leakage or interfering with the model training. Specifically, an adversary can analyze local gradients and the global model to infer clients’ data, and poison local gradients to generate an inaccurate global model. It is extremely challenging to guarantee strong privacy protection of training data while ensuring the robustness of model training. None of the existing studies can achieve the goal. In this paper, we propose a robust privacy-preserving federated learning framework (PILE), which protects the privacy of local gradients and global models, while ensuring their correctness by gradient verification where the server verifies the computation process of local gradients. In PILE, we develop a verifiable perturbation scheme that makes confidential local gradients verifiable for gradient verification. In particular, we build two building blocks of zero-knowledge proofs for the gradient verification without revealing both local gradients and global models. We perform rigorous theoretical analysis that proves the security of PILE and evaluate PILE on both passive and active membership inference attacks. The experiment results show that the attack accuracy under PILE is between $[50.3\%,50.9\%]$ , which is close to the random guesses. Particularly, compared to prior defenses that incur the accuracy losses ranging from 2% to 13%, the accuracy loss of PILE is negligible, i.e., only $\pm 0.3\%$ accuracy loss.
科研通智能强力驱动
Strongly Powered by AbleSci AI