论文标题
延迟意识到半同步客户选择和无线联合学习的模型聚合
Latency Aware Semi-synchronous Client Selection and Model Aggregation for Wireless Federated Learning
论文作者
论文摘要
联合学习(FL)是一个协作机器学习框架,需要不同的客户(例如,物联网设备)通过训练和将其本地模型上传到每个全球迭代中的FL服务器来参与机器学习模型培训过程。从所有客户端接收本地模型后,FL服务器通过汇总收到的本地模型生成全局模型。这种传统的FL过程可能会遇到异质客户设置中的Straggler问题,在该设置中,FL服务器必须等待慢速客户端在每个全球迭代中上传其本地模型,从而增加整体培训时间。解决方案之一是设置截止日期,只有在FL过程中选择截止日期之前可以上传本地模型的客户。由于客户选择有限,该解决方案可能导致趋势速度缓慢和全球模型过度适应问题。在本文中,我们提出了延迟意识到的半同步客户端选择和联合学习方法的模型聚合(课程)方法,该方法允许所有客户端参与整个FL过程,但使用不同的频率。也就是说,比慢速客户端将安排更快的客户端上传模型,从而解决Straggler问题并加速收敛速度,同时避免模型过度拟合。此外,教训能够通过改变截止日期来调整模型准确性和收敛率之间的权衡。进行了广泛的模拟,以将课程的性能与其他两种基线方法(即FedAvg和FedC)进行比较。模拟结果表明,与FedAvg和FedCS相比,课程达到的收敛速度更快,并且模型的准确性比FedCS更高。
Federated learning (FL) is a collaborative machine learning framework that requires different clients (e.g., Internet of Things devices) to participate in the machine learning model training process by training and uploading their local models to an FL server in each global iteration. Upon receiving the local models from all the clients, the FL server generates a global model by aggregating the received local models. This traditional FL process may suffer from the straggler problem in heterogeneous client settings, where the FL server has to wait for slow clients to upload their local models in each global iteration, thus increasing the overall training time. One of the solutions is to set up a deadline and only the clients that can upload their local models before the deadline would be selected in the FL process. This solution may lead to a slow convergence rate and global model overfitting issues due to the limited client selection. In this paper, we propose the Latency awarE Semi-synchronous client Selection and mOdel aggregation for federated learNing (LESSON) method that allows all the clients to participate in the whole FL process but with different frequencies. That is, faster clients would be scheduled to upload their models more frequently than slow clients, thus resolving the straggler problem and accelerating the convergence speed, while avoiding model overfitting. Also, LESSON is capable of adjusting the tradeoff between the model accuracy and convergence rate by varying the deadline. Extensive simulations have been conducted to compare the performance of LESSON with the other two baseline methods, i.e., FedAvg and FedCS. The simulation results demonstrate that LESSON achieves faster convergence speed than FedAvg and FedCS, and higher model accuracy than FedCS.