论文标题

VC的理论解释双血统

VC Theoretical Explanation of Double Descent

论文作者

Lee, Eng Hock, Cherkassky, Vladimir

论文摘要

大型多层神经网络的概括性能越来越兴趣,可以接受训练以达到零训练错误,同时对测试数据良好概括。该制度被称为“第二次下降”,似乎与传统的观点相矛盾,即最佳模型复杂性应反映出不足和过度拟合之间的最佳平衡,即偏见差异权衡。本文介绍了双重下降的VC理论分析,并表明可以通过经典的VC将军范围来充分解释。我们使用多种学习方法的经验结果(例如SVM,最小二乘正方形和多层观察者分类器)使用经验结果来说明用于分类的双重下降进行分类的应用。此外,我们讨论了对深度学习社区中VC理论结果误解的几个原因。

There has been growing interest in generalization performance of large multilayer neural networks that can be trained to achieve zero training error, while generalizing well on test data. This regime is known as 'second descent' and it appears to contradict the conventional view that optimal model complexity should reflect an optimal balance between underfitting and overfitting, i.e., the bias-variance trade-off. This paper presents a VC-theoretical analysis of double descent and shows that it can be fully explained by classical VC-generalization bounds. We illustrate an application of analytic VC-bounds for modeling double descent for classification, using empirical results for several learning methods, such as SVM, Least Squares, and Multilayer Perceptron classifiers. In addition, we discuss several reasons for the misinterpretation of VC-theoretical results in Deep Learning community.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源