论文标题
优化对强大神经网络的信息丢失
Optimizing Information Loss Towards Robust Neural Networks
论文作者
论文摘要
神经网络(NNS)容易受到对抗性例子的影响。这些输入仅与良性的良性略有不同,但引起了对攻击的NNS的错误分类。制作这些例子所需的扰动通常可以忽略不计,甚至无法察觉。为了保护基于深度学习的系统免受此类攻击的侵害,已经提出了一些对抗训练的对策,但仍被认为是最有效的。在这里,使用对抗性示例进行迭代训练,形成一个昂贵且耗时的过程通常会导致性能下降。为了克服对抗性训练的弊端,同时仍提供高水平的安全性,我们提出了一种新的培训方法,我们称为\ textit {entropic retroning}。基于信息理论启发的分析,熵训练模仿对抗训练的影响,而无需努力生成对抗性例子。我们从经验上表明,熵再培训会导致NNS的安全性和鲁棒性显着提高,同时仅依靠给定的原始数据。通过我们的原型实施,我们验证并显示了我们方法对各种NN架构和数据集的有效性。
Neural Networks (NNs) are vulnerable to adversarial examples. Such inputs differ only slightly from their benign counterparts yet provoke misclassifications of the attacked NNs. The required perturbations to craft the examples are often negligible and even human imperceptible. To protect deep learning-based systems from such attacks, several countermeasures have been proposed with adversarial training still being considered the most effective. Here, NNs are iteratively retrained using adversarial examples forming a computational expensive and time consuming process often leading to a performance decrease. To overcome the downsides of adversarial training while still providing a high level of security, we present a new training approach we call \textit{entropic retraining}. Based on an information-theoretic-inspired analysis, entropic retraining mimics the effects of adversarial training without the need of the laborious generation of adversarial examples. We empirically show that entropic retraining leads to a significant increase in NNs' security and robustness while only relying on the given original data. With our prototype implementation we validate and show the effectiveness of our approach for various NN architectures and data sets.