论文标题

过度拟合对抗性强大的深度学习

Overfitting in adversarially robust deep learning

论文作者

Rice, Leslie, Wong, Eric, Kolter, J. Zico

论文摘要

在深度学习中使用过度参数化的网络并尽可能长时间训练这是常见的实践。从理论和经验上讲,有许多研究表明,这种做法令人惊讶地不会过分损害分类器的概括性能。在本文中,我们在经验上研究了经过对抗训练的深网的环境,该现象经过训练,以最大程度地减少最坏情况的对抗性扰动下的损失。我们发现,在多个数据集(SVHN,CIFAR-10,CIFAR-100和IMAGENET)和扰动模型($ \ ell_ \ ell_ \ iffty $和$ \ ell_2 $)上,对训练集的过度拟合确实会在很大程度上损害了良好的性能,从而在很大程度上损害了良好的性能。基于这种观察到的效果,我们表明,在对抗训练时几乎所有最近的算法改进的性能可以通过简单地使用早期停止来匹配。我们还表明,诸如双重下降曲线之类的效果仍然在受过对抗训练的模型中发生,但无法解释观察到的过度拟合。最后,我们研究了多种经典和现代的深度学习补救措施,包括过度拟合,包括正则化和数据增强,并发现在通过早期停止获得的收益中,隔离的方法没有显着改善。可以在https://github.com/locuslab/robust_overfitting上找到所有用于复制实验的代码以及预处理的模型权重和培训日志。

It is common practice in deep learning to use overparameterized networks and train for as long as possible; there are numerous studies that show, both theoretically and empirically, that such practices surprisingly do not unduly harm the generalization performance of the classifier. In this paper, we empirically study this phenomenon in the setting of adversarially trained deep networks, which are trained to minimize the loss under worst-case adversarial perturbations. We find that overfitting to the training set does in fact harm robust performance to a very large degree in adversarially robust training across multiple datasets (SVHN, CIFAR-10, CIFAR-100, and ImageNet) and perturbation models ($\ell_\infty$ and $\ell_2$). Based upon this observed effect, we show that the performance gains of virtually all recent algorithmic improvements upon adversarial training can be matched by simply using early stopping. We also show that effects such as the double descent curve do still occur in adversarially trained models, yet fail to explain the observed overfitting. Finally, we study several classical and modern deep learning remedies for overfitting, including regularization and data augmentation, and find that no approach in isolation improves significantly upon the gains achieved by early stopping. All code for reproducing the experiments as well as pretrained model weights and training logs can be found at https://github.com/locuslab/robust_overfitting.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源