论文标题

自适应服务器学习

Adaptive Serverless Learning

论文作者

Gao, Hongchang, Huang, Heng

论文摘要

随着分布式数据的出现,近年来培训机器学习模型引起了越来越多的关注。在这种制度中,已经提出了许多培训方法,例如分散的SGD。但是,所有现有的分散算法仅关注标准SGD。它可能不适用于某些应用,例如该功能高度稀疏和分类的深度分解机,因此需要自适应训练算法。在本文中,我们提出了一种新型的自适应分散训练方法,该方法可以动态地计算数据。据我们所知,这是第一种自适应分散培训方法。我们的理论结果表明,所提出的算法可以相对于工人数量实现线性加速。此外,为了减少沟通高效的开销,我们进一步提出了一种沟通效率的自适应分散培训方法,这也可以相对于工人数量实现线性加速。最后,对不同任务的广泛实验证实了我们提出的两种方法的有效性。

With the emergence of distributed data, training machine learning models in the serverless manner has attracted increasing attention in recent years. Numerous training approaches have been proposed in this regime, such as decentralized SGD. However, all existing decentralized algorithms only focus on standard SGD. It might not be suitable for some applications, such as deep factorization machine in which the feature is highly sparse and categorical so that the adaptive training algorithm is needed. In this paper, we propose a novel adaptive decentralized training approach, which can compute the learning rate from data dynamically. To the best of our knowledge, this is the first adaptive decentralized training approach. Our theoretical results reveal that the proposed algorithm can achieve linear speedup with respect to the number of workers. Moreover, to reduce the communication-efficient overhead, we further propose a communication-efficient adaptive decentralized training approach, which can also achieve linear speedup with respect to the number of workers. At last, extensive experiments on different tasks have confirmed the effectiveness of our proposed two approaches.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源