论文标题
了解对抗性训练和超越的强大过度拟合
Understanding Robust Overfitting of Adversarial Training and Beyond
论文作者
论文摘要
在深层网络的对抗培训中,强大的过度拟合存在。确切的根本原因仍然尚未完全理解。在这里,我们通过比较\ emph {non-overFit}(弱对手)和\ emph {过拟合}(强烈的对手)对抗训练的数据分布来探索强大过度适应的原因,并观察到由弱对手产生的对抗性数据的分布包含弱对敌人的主要小孔数据。但是,强对手产生的对抗数据更加多样化,分布在大损失数据和小数据上。鉴于这些观察结果,我们进一步设计了数据消融对抗训练,并确定一些不值得对对手强度的小损坏数据会导致在强烈的对手模式下强大的过度拟合。为了缓解这个问题,我们提出\ emph {最小损失约束对抗训练}(MLCAT):在MiniBatch中,我们照常学习大型数据,并采取其他措施来增加小数据的损失。从技术上讲,MLCAT在易于学习以防止强大的过度拟合时会阻碍数据拟合;从哲学上讲,MLCAT反映了将浪费变成宝藏并充分利用每个对抗数据的精神。从算法上讲,我们设计了MLCAT的两个实现,并且广泛的实验表明MLCAT可以消除强大的过度拟合并进一步增强对抗性鲁棒性。
Robust overfitting widely exists in adversarial training of deep networks. The exact underlying reasons for this are still not completely understood. Here, we explore the causes of robust overfitting by comparing the data distribution of \emph{non-overfit} (weak adversary) and \emph{overfitted} (strong adversary) adversarial training, and observe that the distribution of the adversarial data generated by weak adversary mainly contain small-loss data. However, the adversarial data generated by strong adversary is more diversely distributed on the large-loss data and the small-loss data. Given these observations, we further designed data ablation adversarial training and identify that some small-loss data which are not worthy of the adversary strength cause robust overfitting in the strong adversary mode. To relieve this issue, we propose \emph{minimum loss constrained adversarial training} (MLCAT): in a minibatch, we learn large-loss data as usual, and adopt additional measures to increase the loss of the small-loss data. Technically, MLCAT hinders data fitting when they become easy to learn to prevent robust overfitting; philosophically, MLCAT reflects the spirit of turning waste into treasure and making the best use of each adversarial data; algorithmically, we designed two realizations of MLCAT, and extensive experiments demonstrate that MLCAT can eliminate robust overfitting and further boost adversarial robustness.