论文标题

联合和集中机器学习的比较评估

Comparative assessment of federated and centralized machine learning

论文作者

Majeed, Ibrahim Abdul, Kaushik, Sagar, Bardhan, Aniruddha, Tadi, Venkata Siva Kumar, Min, Hwang-Ki, Kumaraguru, Karthikeyan, Muni, Rajasekhara Duvvuru

论文摘要

联合学习(FL)是保留机器学习计划的隐私性,在该计划中,培训发生在跨设备联合的数据中,而不是让它们维持用户隐私。通过使未经培训或受过部分训练的模型直接到达单个设备并使用设备拥有的数据进行本地训练的“设备”来确保这一点,并且服务器汇总了所有部分训练有素的模型学习,以更新全局模型。尽管联合学习设置使用梯度下降中几乎所有模型学习方案,但数据可用性的非IID性质带来了某些特征差异,与集中式方案相比会影响培训。在本文中,我们讨论了影响联合学习培训的各种因素,因为数据的非IID分布性质以及联邦学习方法的固有差异以及典型的集中式梯度下降技术。我们从经验上证明了每个设备样品数量的影响以及输出标签对联合学习的分布。除了通过联合学习寻求的隐私优势外,我们还研究使用联合学习框架时是否具有成本优势。我们表明,当要训练的模型大小不是很大时,联邦学习确实具有成本优势。总而言之,我们需要仔细设计用于性能和成本的模型。

Federated Learning (FL) is a privacy preserving machine learning scheme, where training happens with data federated across devices and not leaving them to sustain user privacy. This is ensured by making the untrained or partially trained models to reach directly the individual devices and getting locally trained "on-device" using the device owned data, and the server aggregating all the partially trained model learnings to update a global model. Although almost all the model learning schemes in the federated learning setup use gradient descent, there are certain characteristic differences brought about by the non-IID nature of the data availability, that affects the training in comparison to the centralized schemes. In this paper, we discuss the various factors that affect the federated learning training, because of the non-IID distributed nature of the data, as well as the inherent differences in the federating learning approach as against the typical centralized gradient descent techniques. We empirically demonstrate the effect of number of samples per device and the distribution of output labels on federated learning. In addition to the privacy advantage we seek through federated learning, we also study if there is a cost advantage while using federated learning frameworks. We show that federated learning does have an advantage in cost when the model sizes to be trained are not reasonably large. All in all, we present the need for careful design of model for both performance and cost.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源