论文标题

内核近似的随机特征:算法,理论及以后的调查

Random Features for Kernel Approximation: A Survey on Algorithms, Theory, and Beyond

论文作者

Liu, Fanghui, Huang, Xiaolin, Chen, Yudong, Suykens, Johan A. K.

论文摘要

随机功能是在大规模问题中加快内核方法的最流行技术之一。相关作品已在2017年的Neurips测试奖和2019年的ICML最佳纸质决赛中得到认可。随机功能的作品迅速增长,因此,希望对此主题进行全面概述,以解释各种算法和理论结果之间的联系。在这项调查中,我们系统地检查了过去十年中随机功能的工作。首先,根据其采样方案,学习程序,降低差异属性及其如何利用培训数据来概述代表性随机算法的动机,特征和贡献。其次,我们回顾了以下关键问题集中的理论结果:需要多少个随机特征,以确保学习估计器的经验/预期风险较高的近似质量或无损失。第三,我们对几个大型基准数据集的流行随机特征算法进行全面评估,并讨论其分类的近似质量和预测性能。最后,我们讨论了随机特征与现代过度参数深度神经网络(DNN)之间的关系,包括在DNN分析中使用高维随机特征以及当前理论和经验结果之间的差距。这项调查可能是对该主题的温和介绍,并且作为有兴趣应用代表性算法并理解各种技术假设的理论结果的从业者的用户指南。我们希望这项调查将有助于讨论该主题中的开放问题,更重要的是阐明了未来的研究方向。

Random features is one of the most popular techniques to speed up kernel methods in large-scale problems. Related works have been recognized by the NeurIPS Test-of-Time award in 2017 and the ICML Best Paper Finalist in 2019. The body of work on random features has grown rapidly, and hence it is desirable to have a comprehensive overview on this topic explaining the connections among various algorithms and theoretical results. In this survey, we systematically review the work on random features from the past ten years. First, the motivations, characteristics and contributions of representative random features based algorithms are summarized according to their sampling schemes, learning procedures, variance reduction properties and how they exploit training data. Second, we review theoretical results that center around the following key question: how many random features are needed to ensure a high approximation quality or no loss in the empirical/expected risks of the learned estimator. Third, we provide a comprehensive evaluation of popular random features based algorithms on several large-scale benchmark datasets and discuss their approximation quality and prediction performance for classification. Last, we discuss the relationship between random features and modern over-parameterized deep neural networks (DNNs), including the use of high dimensional random features in the analysis of DNNs as well as the gaps between current theoretical and empirical results. This survey may serve as a gentle introduction to this topic, and as a users' guide for practitioners interested in applying the representative algorithms and understanding theoretical results under various technical assumptions. We hope that this survey will facilitate discussion on the open problems in this topic, and more importantly, shed light on future research directions.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源