论文标题
CFL:大规模点对点网络中联合学习的集群
CFL: Cluster Federated Learning in Large-scale Peer-to-Peer Networks
论文作者
论文摘要
联邦学习(FL)引发了人们对利用客户本地设备的私人数据的广泛兴趣。但是,FL的参数服务器设置不仅具有较高的带宽要求,而且还提出了数据隐私问题和单点故障。在本文中,我们提出了一种称为CFL的高效和隐私协议,该协议是大规模点对点(P2P)网络中FL的首个细粒度的全球模型培训。与P2P网络中以前的FL不同,CFL汇总了本地模型更新参数,从而提高了大量客户端的通信效率。同样,CFL中的汇总是通过引入身份验证的加密方案以安全的方式执行的,其密钥是通过通过建议的基于投票的基于投票的密钥吊销机制来增强的随机成对密钥方案来建立的。严格的分析表明,CFL保证在两个广泛的威胁模型下的本地模型更新参数的隐私和数据完整性和真实性。更重要的是,提出的钥匙吊销机制可以有效抵抗劫持攻击,从而确保通信键的机密性。 TREC06P和TREC07数据集的巧妙实验表明,由CFL训练的全球模型具有良好的分类精度,模型的概括和快速的收敛速率,并且可以实现系统的辍学率。与P2P网络中FL的第一个全球模型培训方案相比,PPT,CFL提高了43.25%的沟通效率。此外,在计算效率方面,CFL优于PPT。
Federated learning (FL) has sparked extensive interest in exploiting the private data on clients' local devices. However, the parameter server setting of FL not only has high bandwidth requirements, but also poses data privacy issues and a single point of failure. In this paper, we propose an efficient and privacy-preserving protocol, dubbed CFL, which is the first fine-grained global model training for FL in large-scale peer-to-peer (P2P) networks. Unlike previous FL in P2P networks, CFL aggregates local model update parameters hierarchically, which improves the communication efficiency facing large amounts of clients. Also, the aggregation in CFL is performed in a secure manner by introducing the authenticated encryption scheme, whose key is established through a random pairwise key scheme enhanced by a proposed voting-based key revocation mechanism. Rigorous analyses show that CFL guarantees the privacy and data integrity and authenticity of local model update parameters under two widespread threat models. More importantly, the proposed key revocation mechanism can effectively resist hijack attacks, thereby ensuring the confidentiality of the communication keys. Ingenious experiments on the Trec06p and Trec07 datasets show that the global model trained by CFL has good classification accuracy, model generalization, and rapid convergence rate, and the dropout-robustness of the system is achieved. Compared to the first global model training protocol for FL in P2P networks, PPT, CFL improves communication efficiency by 43.25%. Also, CFL outperforms PPT in terms of computational efficiency.