论文标题
私人:以低成本保护甘斯免受会员推理攻击
privGAN: Protecting GANs from membership inference attacks at low cost
论文作者
论文摘要
生成的对抗网络(GAN)使释放合成图像成为共享数据的可行方法,而无需释放原始数据集。已经表明,此类合成数据可用于各种下游任务,例如培训分类器,否则这些任务将需要共享原始数据集。但是,最近的工作表明,GAN模型及其合成生成的数据可用于推断训练集成员资格,该对手可以访问整个数据集和一些辅助信息。当前缓解此问题的方法(例如DPGAN)导致产生的样品质量比原始的非私人gan产生的样本质量较差。在这里,我们开发了一个新的gan架构(Privgan),在该建筑中,发电机不仅要欺骗歧视者,还可以捍卫会员推理攻击。新机制为这种攻击方式提供了保护,同时导致下游表演中可忽略不计。此外,我们的算法已被证明可以明确防止对训练组过度拟合,这解释了我们的保护为何如此有效。本文的主要贡献是:i)我们提出了一种新型的gan架构,可以以隐私的方式生成合成数据,而无需其他超级参数调整和架构选择,ii)我们提供了对私人损失功能的最佳解决方案的理论理解,iii iii iii)我们证明了我们模型对几个白色和黑色数据的有效性,以下是对几个白色和黑色数据的有效性,该数据是bbench的三个bench belch shards,即bench bench senchs,iv ev)与非私人gan相比,Privgan产生的合成图像会导致下游性能的损失可忽略不计。
Generative Adversarial Networks (GANs) have made releasing of synthetic images a viable approach to share data without releasing the original dataset. It has been shown that such synthetic data can be used for a variety of downstream tasks such as training classifiers that would otherwise require the original dataset to be shared. However, recent work has shown that the GAN models and their synthetically generated data can be used to infer the training set membership by an adversary who has access to the entire dataset and some auxiliary information. Current approaches to mitigate this problem (such as DPGAN) lead to dramatically poorer generated sample quality than the original non--private GANs. Here we develop a new GAN architecture (privGAN), where the generator is trained not only to cheat the discriminator but also to defend membership inference attacks. The new mechanism provides protection against this mode of attack while leading to negligible loss in downstream performances. In addition, our algorithm has been shown to explicitly prevent overfitting to the training set, which explains why our protection is so effective. The main contributions of this paper are: i) we propose a novel GAN architecture that can generate synthetic data in a privacy preserving manner without additional hyperparameter tuning and architecture selection, ii) we provide a theoretical understanding of the optimal solution of the privGAN loss function, iii) we demonstrate the effectiveness of our model against several white and black--box attacks on several benchmark datasets, iv) we demonstrate on three common benchmark datasets that synthetic images generated by privGAN lead to negligible loss in downstream performance when compared against non--private GANs.