论文标题
训练的技巧袋
Bag of Tricks for Adversarial Training
论文作者
论文摘要
对抗训练(AT)是促进模型鲁棒性的最有效策略之一。但是,最近的基准表明,大多数提议的改进AT不如简单地停止培训程序的效率。这个违反直觉的事实促使我们研究了数十个AT方法的实施细节。令人惊讶的是,我们发现这些方法中使用的基本设置(例如,重量衰减,培训时间表等)非常不一致。在这项工作中,我们对CIFAR-10进行了全面的评估,重点介绍了大多数被忽视的训练技巧和超级参数对对抗训练的模型的影响。我们的经验观察表明,对抗性鲁棒性对某些基本培训环境的敏感性比我们想象的要敏感得多。例如,权重衰减值略有不同可以将模型的鲁棒精度降低超过7%,这可能覆盖所提出的方法引起的潜在促销。我们总结了基线训练设置和重新实现先前的防御能力,以实现新的最新结果。这些事实还吸引了对防御措施进行基准测试时对被忽视的混杂因素的更多担忧。
Adversarial training (AT) is one of the most effective strategies for promoting model robustness. However, recent benchmarks show that most of the proposed improvements on AT are less effective than simply early stopping the training procedure. This counter-intuitive fact motivates us to investigate the implementation details of tens of AT methods. Surprisingly, we find that the basic settings (e.g., weight decay, training schedule, etc.) used in these methods are highly inconsistent. In this work, we provide comprehensive evaluations on CIFAR-10, focusing on the effects of mostly overlooked training tricks and hyperparameters for adversarially trained models. Our empirical observations suggest that adversarial robustness is much more sensitive to some basic training settings than we thought. For example, a slightly different value of weight decay can reduce the model robust accuracy by more than 7%, which is probable to override the potential promotion induced by the proposed methods. We conclude a baseline training setting and re-implement previous defenses to achieve new state-of-the-art results. These facts also appeal to more concerns on the overlooked confounders when benchmarking defenses.