论文标题

超越跨透明镜:学习高度可分离的特征分布,以进行鲁棒和准确的分类

Beyond cross-entropy: learning highly separable feature distributions for robust and accurate classification

论文作者

Ali, Arslan, Migliorati, Andrea, Bianchi, Tiziano, Magli, Enrico

论文摘要

深度学习在包括图像分类在内的多种应用中表现出了出色的性能。但是,已知深层分类器高度容易受到对抗攻击的影响,因为输入的次要扰动很容易导致错误。为对抗性攻击提供鲁棒性是一项非常具有挑战性的任务,尤其是在涉及大量课程的问题中,因为它通常以准确性下降为代价。在这项工作中,我们提出了高斯班级条件单纯形(GCC)损失:一种训练深功能强大的多类分类器的新方法,可提供对抗性鲁棒性,同时实现甚至超过最新方法的分类精度。与其他框架不同,所提出的方法将输入类映射到潜在空间中的目标分布上,以使类是可分离的。我们的目标函数不是最大程度地提高目标标签的目标标签,而是推动网络产生特征分布,从而产生高层间的分离。分布的平均值集中在单纯形的顶点上,使每个类都与其他每个类别相同。我们表明,基于我们方法的潜在空间的正规化可产生出色的分类准确性,并固有地为多种对抗性攻击提供了稳健性,无论是针对性和未靶向,优于挑战性数据集的最先进方法。

Deep learning has shown outstanding performance in several applications including image classification. However, deep classifiers are known to be highly vulnerable to adversarial attacks, in that a minor perturbation of the input can easily lead to an error. Providing robustness to adversarial attacks is a very challenging task especially in problems involving a large number of classes, as it typically comes at the expense of an accuracy decrease. In this work, we propose the Gaussian class-conditional simplex (GCCS) loss: a novel approach for training deep robust multiclass classifiers that provides adversarial robustness while at the same time achieving or even surpassing the classification accuracy of state-of-the-art methods. Differently from other frameworks, the proposed method learns a mapping of the input classes onto target distributions in a latent space such that the classes are linearly separable. Instead of maximizing the likelihood of target labels for individual samples, our objective function pushes the network to produce feature distributions yielding high inter-class separation. The mean values of the distributions are centered on the vertices of a simplex such that each class is at the same distance from every other class. We show that the regularization of the latent space based on our approach yields excellent classification accuracy and inherently provides robustness to multiple adversarial attacks, both targeted and untargeted, outperforming state-of-the-art approaches over challenging datasets.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源