论文标题
涂鸦:贝叶斯生成模型的无梯度联合学习
GRAFFL: Gradient-free Federated Learning of a Bayesian Generative Model
论文作者
论文摘要
联合学习平台正在越来越受欢迎。主要好处之一是减轻隐私风险,因为可以在不收集或共享数据的情况下实现算法的学习。尽管联盟学习(即许多基于随机梯度算法)表现出巨大的希望,但在保护隐私方面仍然存在许多挑战性的问题,尤其是在渐变更新和交换过程中。本文介绍了第一个无梯度的联合学习框架,称为Graffl,用于学习基于近似贝叶斯计算的贝叶斯生成模型。与基于梯度的常规联合学习算法不同,我们的框架不需要拆卸模型(即,到线性组件)或浏览数据(或汇总数据加密)以保留隐私。相反,该框架使用从每个参与机构得出的隐式信息来学习参数的后验分布。隐式信息是摘要统计数据,该统计数据是从本研究中开发的一个神经网络,旨在创建压缩和线性分离的表示形式,从而保护敏感信息免受泄漏的影响。作为一种足够的维度降低技术,这被证明提供了足够的汇总统计数据。我们建议基于涂鸦的贝叶斯高斯混合模型作为框架的概念验证。使用几个数据集,我们在隐私保护和预测性能(即接近理想的设置)方面证明了模型的可行性和实用性。训练有素的模型作为准全球模型可以生成涉及其他机构信息的信息样本,并增强每个机构的数据分析。
Federated learning platforms are gaining popularity. One of the major benefits is to mitigate the privacy risks as the learning of algorithms can be achieved without collecting or sharing data. While federated learning (i.e., many based on stochastic gradient algorithms) has shown great promise, there are still many challenging problems in protecting privacy, especially during the process of gradients update and exchange. This paper presents the first gradient-free federated learning framework called GRAFFL for learning a Bayesian generative model based on approximate Bayesian computation. Unlike conventional federated learning algorithms based on gradients, our framework does not require to disassemble a model (i.e., to linear components) or to perturb data (or encryption of data for aggregation) to preserve privacy. Instead, this framework uses implicit information derived from each participating institution to learn posterior distributions of parameters. The implicit information is summary statistics derived from SuffiAE that is a neural network developed in this study to create compressed and linearly separable representations thereby protecting sensitive information from leakage. As a sufficient dimensionality reduction technique, this is proved to provide sufficient summary statistics. We propose the GRAFFL-based Bayesian Gaussian mixture model to serve as a proof-of-concept of the framework. Using several datasets, we demonstrated the feasibility and usefulness of our model in terms of privacy protection and prediction performance (i.e., close to an ideal setting). The trained model as a quasi-global model can generate informative samples involving information from other institutions and enhances data analysis of each institution.