论文标题

Perdoor:使用对抗性扰动的联合学习中的持续性不均匀后门

PerDoor: Persistent Non-Uniform Backdoors in Federated Learning using Adversarial Perturbations

论文作者

Alam, Manaar, Sarkar, Esha, Maniatakos, Michail

论文摘要

联合学习(FL)使众多参与者能够在不暴露其个人,潜在敏感的数据的情况下协作培训深度学习模型,从而使其成为协作培训中数据隐私的有前途的解决方案。但是,FL和未经审查的数据的分布式性质使其本质上容易受到后门攻击的影响:在这种情况下,对手在训练过程中将后门功能注入了集中式模型,这可能会触发,从而导致所需的错误分类对于特定的对手选择的输入。一系列先前的工作在FL系统中建立了成功的后门注入。但是,这些后门没有被证明是持久的。如果从训练过程中删除对手,则后门功能不会保留在系统中,因为在连续的FL训练回合中,集中式模型参数连续突变。因此,在这项工作中,我们提出了Perdoor,这是FL的持续逐型后门注入技术,这是由对抗性扰动和靶向集中模型的靶向参数驱动的,在连续的FL弹中偏离较小,对主要任务准确性的贡献最小。考虑到图像分类方案的详尽评估,与传统的后门攻击相比,在多个FL的情况下,平均$ 10.5 \ times $持续存在。通过实验,我们进一步在FL系统中存在最新的后门预防技术的情况下进一步表现出了多门的效力。此外,与现有的后门技术的均匀触发器(具有固定的模式和位置​​)相比,对抗扰动的操作还有助于为户外开发用于后门输入的非均匀触发模式,这些触发器(具有固定的模式和位置​​),这些技术容易容易缓解。

Federated Learning (FL) enables numerous participants to train deep learning models collaboratively without exposing their personal, potentially sensitive data, making it a promising solution for data privacy in collaborative training. The distributed nature of FL and unvetted data, however, makes it inherently vulnerable to backdoor attacks: In this scenario, an adversary injects backdoor functionality into the centralized model during training, which can be triggered to cause the desired misclassification for a specific adversary-chosen input. A range of prior work establishes successful backdoor injection in an FL system; however, these backdoors are not demonstrated to be long-lasting. The backdoor functionality does not remain in the system if the adversary is removed from the training process since the centralized model parameters continuously mutate during successive FL training rounds. Therefore, in this work, we propose PerDoor, a persistent-by-construction backdoor injection technique for FL, driven by adversarial perturbation and targeting parameters of the centralized model that deviate less in successive FL rounds and contribute the least to the main task accuracy. An exhaustive evaluation considering an image classification scenario portrays on average $10.5\times$ persistence over multiple FL rounds compared to traditional backdoor attacks. Through experiments, we further exhibit the potency of PerDoor in the presence of state-of-the-art backdoor prevention techniques in an FL system. Additionally, the operation of adversarial perturbation also assists PerDoor in developing non-uniform trigger patterns for backdoor inputs compared to uniform triggers (with fixed patterns and locations) of existing backdoor techniques, which are prone to be easily mitigated.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源