论文标题
垂直联合学习中DNN培训的功能重建攻击和对策
Feature Reconstruction Attacks and Countermeasures of DNN training in Vertical Federated Learning
论文作者
论文摘要
联邦学习(FL)越来越多地以其垂直形式部署,以促进对孤立数据的安全协作培训。在垂直FL(VFL)中,参与者拥有同一组样本实例的不相交特征。其中,只有一个标签。该参与者被称为主动党,启动培训并与其他参与者互动,称为被动党。尽管VFL采用的采用越来越高,但在很大程度上仍然未知是否以及如何从被动方提取特征数据,尤其是在训练深神经网络(DNN)模型时。 本文首次尝试研究了VFL中DNN培训的功能安全问题。我们考虑了在主动方和被动方之间分区的DNN模型,其中后者仅包含输入层的一个子集,并显示了二进制值的某些类别特征。使用确切的封面问题减少,我们证明重建这些二进制特征是NP-HARD。通过分析,我们证明,除非特征维度极大,否则在理论上和实际上,都可以使用高效的基于搜索的算法来发起重建攻击,该算法比当前功能保护技术占上风。为了解决这个问题,我们针对重建攻击开发了一种新颖的功能保护方案,该方案有效地误导了搜索到某些预先指定的随机值。通过广泛的实验,我们表明我们的保护方案在各种VFL应用中维持特征重建攻击,而无需精确损失。
Federated learning (FL) has increasingly been deployed, in its vertical form, among organizations to facilitate secure collaborative training over siloed data. In vertical FL (VFL), participants hold disjoint features of the same set of sample instances. Among them, only one has labels. This participant, known as the active party, initiates the training and interacts with the other participants, known as the passive parties. Despite the increasing adoption of VFL, it remains largely unknown if and how the active party can extract feature data from the passive party, especially when training deep neural network (DNN) models. This paper makes the first attempt to study the feature security problem of DNN training in VFL. We consider a DNN model partitioned between active and passive parties, where the latter only holds a subset of the input layer and exhibits some categorical features of binary values. Using a reduction from the Exact Cover problem, we prove that reconstructing those binary features is NP-hard. Through analysis, we demonstrate that, unless the feature dimension is exceedingly large, it remains feasible, both theoretically and practically, to launch a reconstruction attack with an efficient search-based algorithm that prevails over current feature protection techniques. To address this problem, we develop a novel feature protection scheme against the reconstruction attack that effectively misleads the search to some pre-specified random values. With an extensive set of experiments, we show that our protection scheme sustains the feature reconstruction attack in various VFL applications at no expense of accuracy loss.