论文标题
通过异常检测确定联邦学习中的后门攻击
Identifying Backdoor Attacks in Federated Learning via Anomaly Detection
论文作者
论文摘要
近年来,近年来,联邦学习的收养量增加了,以应对日益增长的数据隐私需求。但是,联邦学习的不透明的本地培训过程也引发了人们对模型忠诚的关注。例如,研究表明,联邦学习容易受到后门攻击的影响,因此,受损的参与者可以在存在后门触发器的情况下偷偷地修改模型的行为。本文通过检查共享模型更新提出了有效的防御攻击辩护。我们首先要观察到,后门的嵌入会影响参与者的本地模型权重,从其模型梯度的幅度和方向来影响,这可能表现为可区分的差异。我们通过研究模型的梯度子集的统计分布来实现对后门的强大识别。具体而言,我们首先将模型梯度分为片段向量,该元素代表模型参数的一小部分。然后,我们采用异常检测来定位分布偏斜的片段,并以最大离群值修剪参与者。我们在一种新颖的防御方法阿里巴(Ariba)中体现了发现。我们通过广泛的分析表明,我们提出的方法有效地减轻了最新的后门攻击,对任务实用程序的影响很小。
Federated learning has seen increased adoption in recent years in response to the growing regulatory demand for data privacy. However, the opaque local training process of federated learning also sparks rising concerns about model faithfulness. For instance, studies have revealed that federated learning is vulnerable to backdoor attacks, whereby a compromised participant can stealthily modify the model's behavior in the presence of backdoor triggers. This paper proposes an effective defense against the attack by examining shared model updates. We begin with the observation that the embedding of backdoors influences the participants' local model weights in terms of the magnitude and orientation of their model gradients, which can manifest as distinguishable disparities. We enable a robust identification of backdoors by studying the statistical distribution of the models' subsets of gradients. Concretely, we first segment the model gradients into fragment vectors that represent small portions of model parameters. We then employ anomaly detection to locate the distributionally skewed fragments and prune the participants with the most outliers. We embody the findings in a novel defense method, ARIBA. We demonstrate through extensive analyses that our proposed methods effectively mitigate state-of-the-art backdoor attacks with minimal impact on task utility.