论文标题

亚种群数据中毒攻击

Subpopulation Data Poisoning Attacks

论文作者

Jagielski, Matthew, Severi, Giorgio, Harger, Niklas Pousette, Oprea, Alina

论文摘要

机器学习系统部署在关键设置中,但它们可能以意外的方式失败,从而影响其预测的准确性。对机器学习的中毒攻击会引起机器学习算法使用的数据的对抗性修改,以选择性地更改其输出。在这项工作中,我们引入了一种新型的数据中毒攻击,称为\ emph {亚群攻击},当数据集大而多样化时,这尤其重要。我们设计了用于亚群攻击的模块化框架,用不同的构建块对其进行实例化,并表明攻击对于各种数据集和机器学习模型有效。我们使用影响功能和梯度优化方法进一步优化连续域中的攻击。与现有的后门中毒攻击相比,亚群攻击的优势是在推理时间在自然分布的数据点中诱导错误分类,从而使攻击非常隐秘。我们还表明,我们的攻击策略可用于改善现有的目标攻击。我们证明,在某些假设下,亚种群攻击是不可能防御的,并从经验上证明了现有防御措施抵抗我们的攻击的局限性,强调了保护机器学习免受这种威胁的困难。

Machine learning systems are deployed in critical settings, but they might fail in unexpected ways, impacting the accuracy of their predictions. Poisoning attacks against machine learning induce adversarial modification of data used by a machine learning algorithm to selectively change its output when it is deployed. In this work, we introduce a novel data poisoning attack called a \emph{subpopulation attack}, which is particularly relevant when datasets are large and diverse. We design a modular framework for subpopulation attacks, instantiate it with different building blocks, and show that the attacks are effective for a variety of datasets and machine learning models. We further optimize the attacks in continuous domains using influence functions and gradient optimization methods. Compared to existing backdoor poisoning attacks, subpopulation attacks have the advantage of inducing misclassification in naturally distributed data points at inference time, making the attacks extremely stealthy. We also show that our attack strategy can be used to improve upon existing targeted attacks. We prove that, under some assumptions, subpopulation attacks are impossible to defend against, and empirically demonstrate the limitations of existing defenses against our attacks, highlighting the difficulty of protecting machine learning against this threat.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源