论文标题
审核私人机器学习的一般框架
A General Framework for Auditing Differentially Private Machine Learning
论文作者
论文摘要
我们提出了一个框架,以统计审核差异化机器学习者在实践中授予的隐私保证。尽管以前的作品已经采取了通过中毒攻击或会员推理来评估隐私损失的步骤,但它们是针对特定模型量身定制的,或者表现出低统计能力。我们的工作开发了一种一般方法,以经验评估差异化机器学习实现的隐私,将改进的隐私搜索和验证方法与基于影响力的中毒攻击的工具包相结合。我们证明了对先前方法的审计能力显着提高了各种模型,包括逻辑回归,天真的贝叶斯和随机森林。我们的方法可用于检测由于实施错误或滥用而导致的侵犯隐私行为。如果不存在违规行为,它可以帮助理解可以从给定数据集,算法和隐私规范中泄漏的信息量。
We present a framework to statistically audit the privacy guarantee conferred by a differentially private machine learner in practice. While previous works have taken steps toward evaluating privacy loss through poisoning attacks or membership inference, they have been tailored to specific models or have demonstrated low statistical power. Our work develops a general methodology to empirically evaluate the privacy of differentially private machine learning implementations, combining improved privacy search and verification methods with a toolkit of influence-based poisoning attacks. We demonstrate significantly improved auditing power over previous approaches on a variety of models including logistic regression, Naive Bayes, and random forest. Our method can be used to detect privacy violations due to implementation errors or misuse. When violations are not present, it can aid in understanding the amount of information that can be leaked from a given dataset, algorithm, and privacy specification.