论文标题

基于数据增强的防御方法,反对神经网络中的对抗攻击

A Data Augmentation-based Defense Method Against Adversarial Attacks in Neural Networks

论文作者

Zeng, Yi, Qiu, Han, Memmi, Gerard, Qiu, Meikang

论文摘要

众所周知,计算机视觉(CV)中的深神经网络(DNN)很容易受到对抗性示例(AES)的影响,即恶意添加了恶意的扰动,以引起错误的分类结果。这种可变性一直是现实生活中DNN作为核心组成部分的系统的潜在风险。关于如何保护DNN模型免受AES解决的研究已经进行了许多努力。但是,以前的工作无法有效地减少新型对抗性攻击所引起的影响,并同时与现实生活中的约束兼容。在本文中,我们专注于开发一种轻巧的防御方法,该方法可以有效地使全白盒对抗性攻击与现实生活约束的兼容性无效。从基本的仿射变换中,我们将三个转换与随机系数整合在一起,以微调尊重辩护样品的变化量。与过去两年中顶级AI会议上发表的4种最先进的防御方法相比,我们的方法表现出了出色的鲁棒性和效率。值得一提的是,我们的模型可以承受高级自适应攻击,即BPDA,并以50发子弹为基础,并且仍然可以帮助目标模型保持精度约80%,同时将攻击成功率限制在几乎零。

Deep Neural Networks (DNNs) in Computer Vision (CV) are well-known to be vulnerable to Adversarial Examples (AEs), namely imperceptible perturbations added maliciously to cause wrong classification results. Such variability has been a potential risk for systems in real-life equipped DNNs as core components. Numerous efforts have been put into research on how to protect DNN models from being tackled by AEs. However, no previous work can efficiently reduce the effects caused by novel adversarial attacks and be compatible with real-life constraints at the same time. In this paper, we focus on developing a lightweight defense method that can efficiently invalidate full whitebox adversarial attacks with the compatibility of real-life constraints. From basic affine transformations, we integrate three transformations with randomized coefficients that fine-tuned respecting the amount of change to the defended sample. Comparing to 4 state-of-art defense methods published in top-tier AI conferences in the past two years, our method demonstrates outstanding robustness and efficiency. It is worth highlighting that, our model can withstand advanced adaptive attack, namely BPDA with 50 rounds, and still helps the target model maintain an accuracy around 80 %, meanwhile constraining the attack success rate to almost zero.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源