论文标题
FEDCOS:一种场景自适应联合优化的增强,以提高性能
FedCos: A Scene-adaptive Federated Optimization Enhancement for Performance Improvement
论文作者
论文摘要
作为一项新兴技术,联邦学习(FL)涉及分布式边缘设备上的培训机器学习模型,这引起了人们的关注,并且已经进行了广泛的研究。但是,与集中式培训相比,客户数据的异质性严重降低了FL的性能。它导致当地训练的客户模型朝不同的方向移动。一方面,它会减慢甚至失速全局更新,从而导致通信效率低下。另一方面,它扩大了本地模型之间的距离,从而导致了汇总的全球模型,其性能差。幸运的是,可以通过减少本地模型移动的方向之间的角度来减轻这些缺点。根据这一事实,我们提出了FedCos,从而通过引入余弦相似性惩罚来降低局部模型的定向不一致。它促进了朝向辅助全球方向的本地模型迭代。此外,我们的方法是自动适应各种非IID设置,而无需精心选择超参数。实验结果表明,FEDCO的表现优于众所周知的基线,可以在各种FL场景下增强它们,包括不同程度的数据异质性,不同数量的参与者以及跨索洛和交叉设置设置。此外,FEDCO将沟通效率提高了2至5次。借助FedCos,多种FL方法所需的通信回合要比以前要少得多,以获得具有可比性能的模型。
As an emerging technology, federated learning (FL) involves training machine learning models over distributed edge devices, which attracts sustained attention and has been extensively studied. However, the heterogeneity of client data severely degrades the performance of FL compared with that in centralized training. It causes the locally trained models of clients to move in different directions. On the one hand, it slows down or even stalls the global updates, leading to inefficient communication. On the other hand, it enlarges the distances between local models, resulting in an aggregated global model with poor performance. Fortunately, these shortcomings can be mitigated by reducing the angle between the directions that local models move in. Based on this fact, we propose FedCos, which reduces the directional inconsistency of local models by introducing a cosine-similarity penalty. It promotes the local model iterations towards an auxiliary global direction. Moreover, our approach is auto-adapt to various non-IID settings without an elaborate selection of hyperparameters. The experimental results show that FedCos outperforms the well-known baselines and can enhance them under a variety of FL scenes, including varying degrees of data heterogeneity, different number of participants, and cross-silo and cross-device settings. Besides, FedCos improves communication efficiency by 2 to 5 times. With the help of FedCos, multiple FL methods require significantly fewer communication rounds than before to obtain a model with comparable performance.