论文标题
多空格:使用机器学习扩展针对网络钓鱼网站检测器的对抗性攻击的逃避空间
Multi-SpacePhish: Extending the Evasion-space of Adversarial Attacks against Phishing Website Detectors using Machine Learning
论文作者
论文摘要
现有有关对抗机器学习(ML)的文献重点是展示打破所有ML模型的攻击,或者要承受大多数攻击的防御措施。不幸的是,几乎没有考虑到袭击或辩护的实际可行性。此外,对抗样本通常在“功能空间”中制作,从而对可疑价值进行相应的评估。简而言之,当前情况不允许估计对抗性攻击构成的实际威胁,从而导致缺乏安全的ML系统。 我们的目的是在本文中阐明这种混乱。通过考虑ML在网络钓鱼网站检测(PWD)中的应用,我们可以将“逃避空间”形式化,可以引入对抗性扰动来欺骗ML-PWD - 证明即使在“功能空间”中的扰动也很有用。然后,我们提出了一个现实的威胁模型,描述了针对ML-PWD的逃避攻击,该模型的舞台价格便宜,因此对真正的教育师而言,本质上更具吸引力。之后,我们对12次逃避攻击进行了首次对最先进的ML-PWD进行统计验证的评估。我们的评估表明(i)更可能发生逃避尝试的真正功效; (ii)在不同的逃避空间中精心制作的扰动的影响。我们现实的逃避尝试引起了统计学上的显着降解(p <0.05时为3-10%),它们的廉价成本使它们成为微妙的威胁。但是,值得注意的是,某些ML-PWD不受我们最现实的攻击的影响(p = 0.22)。 最后,作为该期刊出版物的另一个贡献,我们是第一个考虑一个有趣的案例,即攻击者同时在多个逃避空间中引入扰动。这些新结果表明,在问题和功能空间中同时应用扰动可能会导致检测率从0.95下降到0。
Existing literature on adversarial Machine Learning (ML) focuses either on showing attacks that break every ML model, or defenses that withstand most attacks. Unfortunately, little consideration is given to the actual feasibility of the attack or the defense. Moreover, adversarial samples are often crafted in the "feature-space", making the corresponding evaluations of questionable value. Simply put, the current situation does not allow to estimate the actual threat posed by adversarial attacks, leading to a lack of secure ML systems. We aim to clarify such confusion in this paper. By considering the application of ML for Phishing Website Detection (PWD), we formalize the "evasion-space" in which an adversarial perturbation can be introduced to fool a ML-PWD -- demonstrating that even perturbations in the "feature-space" are useful. Then, we propose a realistic threat model describing evasion attacks against ML-PWD that are cheap to stage, and hence intrinsically more attractive for real phishers. After that, we perform the first statistically validated assessment of state-of-the-art ML-PWD against 12 evasion attacks. Our evaluation shows (i) the true efficacy of evasion attempts that are more likely to occur; and (ii) the impact of perturbations crafted in different evasion-spaces. Our realistic evasion attempts induce a statistically significant degradation (3-10% at p<0.05), and their cheap cost makes them a subtle threat. Notably, however, some ML-PWD are immune to our most realistic attacks (p=0.22). Finally, as an additional contribution of this journal publication, we are the first to consider the intriguing case wherein an attacker introduces perturbations in multiple evasion-spaces at the same time. These new results show that simultaneously applying perturbations in the problem- and feature-space can cause a drop in the detection rate from 0.95 to 0.