论文标题
RESSFL:用于防御模型反转攻击的电阻转移框架
ResSFL: A Resistance Transfer Framework for Defending Model Inversion Attack in Split Federated Learning
论文作者
论文摘要
这项工作旨在解决对分裂联邦学习(SFL)的模型反转(MI)攻击。 SFL是一种最近的分布式培训方案,其中多个客户将中间激活(即功能映射)而不是原始数据发送到中央服务器。尽管这样的方案有助于减少客户端端的计算负载,但它可以自身从服务器中间激活中重建原始数据。现有的保护SFL仅考虑推论,并且在训练过程中不处理攻击。因此,我们提出了Ressfl,这是一个分裂的联邦学习框架,旨在在培训期间具有耐药性。它基于通过攻击者感知训练得出抗性功能提取器,并在标准SFL培训之前使用此提取器来初始化客户端模型。这种方法有助于降低由于在客户端对抗训练中使用强反转模型以及在早期训练时期发起的攻击的脆弱性而导致的计算复杂性。在CIFAR-100数据集上,我们提出的框架成功地减轻了MI攻击对VGG-11模型的攻击,其均值均值均值为0.050,而基线系统获得的0.005。该框架的精度为67.5%(仅1%的精度下降),其计算开销非常低。代码在以下网址发布:https://github.com/zlijingtao/ressfl。
This work aims to tackle Model Inversion (MI) attack on Split Federated Learning (SFL). SFL is a recent distributed training scheme where multiple clients send intermediate activations (i.e., feature map), instead of raw data, to a central server. While such a scheme helps reduce the computational load at the client end, it opens itself to reconstruction of raw data from intermediate activation by the server. Existing works on protecting SFL only consider inference and do not handle attacks during training. So we propose ResSFL, a Split Federated Learning Framework that is designed to be MI-resistant during training. It is based on deriving a resistant feature extractor via attacker-aware training, and using this extractor to initialize the client-side model prior to standard SFL training. Such a method helps in reducing the computational complexity due to use of strong inversion model in client-side adversarial training as well as vulnerability of attacks launched in early training epochs. On CIFAR-100 dataset, our proposed framework successfully mitigates MI attack on a VGG-11 model with a high reconstruction Mean-Square-Error of 0.050 compared to 0.005 obtained by the baseline system. The framework achieves 67.5% accuracy (only 1% accuracy drop) with very low computation overhead. Code is released at: https://github.com/zlijingtao/ResSFL.