论文标题
探索参数有效的微调以实现联合学习的基础模型
Exploring Parameter-Efficient Fine-Tuning to Enable Foundation Models in Federated Learning
论文作者
论文摘要
联邦学习(FL)已成为一种有希望的范式,用于实现模型的协作培训,而无需集中访问本地设备上的原始数据。在典型的FL范式(例如,FedAvg)中,每轮型号都会向服务器发送到参与的客户。最近,使用小型预训练模型的使用已被证明可以有效地使用联合学习优化和改善收敛性。但是,最近最新的预培训模型变得越来越有能力,但也具有更多的参数,称为“基础模型”。在常规的FL中,共享巨大的模型权重可以迅速给系统带来巨大的通信负担,尤其是在使用更多功能的模型的情况下。我们能否找到一种解决方案,以使FL中的那些强大且随时可用的预训练模型同时减轻沟通负担,以实现出色的性能?为此,我们研究了在联合学习中使用参数有效调整的使用,因此引入了一个新框架:fedepeft。具体而言,我们从系统地评估了FedPeft在各种客户端稳定性,数据分配和差异隐私设置中的性能。通过仅在本地调整和全球共享模型权重的一小部分,可以在整个联合学习的方案中保持竞争性甚至更好的表现,从而实现总沟通开销的大量降低,从而为实用和有效的联合系统提供了新的范式,从而为新的范式提供了洞察力。
Federated learning (FL) has emerged as a promising paradigm for enabling the collaborative training of models without centralized access to the raw data on local devices. In the typical FL paradigm (e.g., FedAvg), model weights are sent to and from the server each round to participating clients. Recently, the use of small pre-trained models has been shown to be effective in federated learning optimization and improving convergence. However, recent state-of-the-art pre-trained models are getting more capable but also have more parameters, known as the "Foundation Models." In conventional FL, sharing the enormous model weights can quickly put a massive communication burden on the system, especially if more capable models are employed. Can we find a solution to enable those strong and readily available pre-trained models in FL to achieve excellent performance while simultaneously reducing the communication burden? To this end, we investigate the use of parameter-efficient fine-tuning in federated learning and thus introduce a new framework: FedPEFT. Specifically, we systemically evaluate the performance of FedPEFT across a variety of client stability, data distribution, and differential privacy settings. By only locally tuning and globally sharing a small portion of the model weights, significant reductions in the total communication overhead can be achieved while maintaining competitive or even better performance in a wide range of federated learning scenarios, providing insight into a new paradigm for practical and effective federated systems.