论文标题

实用的无盒对抗攻击,没有训练的混合图像转换

Practical No-box Adversarial Attacks with Training-free Hybrid Image Transformation

论文作者

Zhang, Qilong, Sun, Youheng, Zhang, Chaoning, Li, Chaoqun, Wang, Xuanhan, Song, Jingkuan, Gao, Lianli

论文摘要

近年来,深神经网络(DNN)的对抗脆弱性引起了人们的关注。在所有威胁模型中,无箱攻击是最实用但极具挑战性的,因为它们既不依赖于目标模型或类似替代模型的知识,也没有访问数据集以培训新的替代模型。尽管最近的一种方法试图从宽松的意义上尝试这种攻击,但其性能还不够好,培训的计算开销很昂贵。在本文中,我们向前迈进一步,并展示No-box威胁模型下的\ textbf {triage-free}对抗扰动的存在,该模型可以成功地实时攻击不同的DNN。通过观察到低级特征的高频组件(HFC)域的动机,并且在分类中起着至关重要的作用,我们主要通过操纵其频率成分来攻击图像。具体而言,通过抑制原始的HFC和嘈杂的HFC来操纵扰动。我们通过经验和实验分析有效的嘈杂HFC的要求,并表明它应该是区域均匀的,重复和致密的。 ImageNet数据集的广泛实验证明了我们提出的NO-Box方法的有效性。它平均攻击了十个成功率\ textbf {98.13 \%}的著名模型,从\ textbf {29.39 \%})均超过了最先进的no-box攻击。此外,我们的方法甚至在基于主流传输的黑盒攻击方面具有竞争力。

In recent years, the adversarial vulnerability of deep neural networks (DNNs) has raised increasing attention. Among all the threat models, no-box attacks are the most practical but extremely challenging since they neither rely on any knowledge of the target model or similar substitute model, nor access the dataset for training a new substitute model. Although a recent method has attempted such an attack in a loose sense, its performance is not good enough and computational overhead of training is expensive. In this paper, we move a step forward and show the existence of a \textbf{training-free} adversarial perturbation under the no-box threat model, which can be successfully used to attack different DNNs in real-time. Motivated by our observation that high-frequency component (HFC) domains in low-level features and plays a crucial role in classification, we attack an image mainly by manipulating its frequency components. Specifically, the perturbation is manipulated by suppression of the original HFC and adding of noisy HFC. We empirically and experimentally analyze the requirements of effective noisy HFC and show that it should be regionally homogeneous, repeating and dense. Extensive experiments on the ImageNet dataset demonstrate the effectiveness of our proposed no-box method. It attacks ten well-known models with a success rate of \textbf{98.13\%} on average, which outperforms state-of-the-art no-box attacks by \textbf{29.39\%}. Furthermore, our method is even competitive to mainstream transfer-based black-box attacks.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源