论文标题

通过重新聚合验证ML系统

Verification of ML Systems via Reparameterization

论文作者

Tristan, Jean-Baptiste, Tassarotti, Joseph, Vajjha, Koundinya, Wick, Michael L., Banerjee, Anindya

论文摘要

随着机器学习越来越多地用于基本系统,重要的是减少或消除严重错误的发生率。越来越多的研究开发了机器学习算法,并正式保证了性能,稳健性或公平性。然而,对这些算法的分析通常很复杂,实践中实现此类系统会引入错误空间。证明助手可用于通过构造机器检查正确性的证明来正式验证机器学习系统,以排除此类错误。但是,关于证明助手内部概率主张的推理仍然具有挑战性。我们展示了如何使用\ emph {Reparameterization}的概念在定理宣传中自动表示概率程序,以及如何从概率程序中自动生成一些乏味的可测量性证明。为了证明这种方法足够广泛,可以处理相当不同的机器学习系统,我们验证了统计学习理论的经典结果(决策树桩的PAC-可学习性),并证明了贝叶斯假设检验中使用的无效模型满足了公平的标准,称为人口统计学奇偶。

As machine learning is increasingly used in essential systems, it is important to reduce or eliminate the incidence of serious bugs. A growing body of research has developed machine learning algorithms with formal guarantees about performance, robustness, or fairness. Yet, the analysis of these algorithms is often complex, and implementing such systems in practice introduces room for error. Proof assistants can be used to formally verify machine learning systems by constructing machine checked proofs of correctness that rule out such bugs. However, reasoning about probabilistic claims inside of a proof assistant remains challenging. We show how a probabilistic program can be automatically represented in a theorem prover using the concept of \emph{reparameterization}, and how some of the tedious proofs of measurability can be generated automatically from the probabilistic program. To demonstrate that this approach is broad enough to handle rather different types of machine learning systems, we verify both a classic result from statistical learning theory (PAC-learnability of decision stumps) and prove that the null model used in a Bayesian hypothesis test satisfies a fairness criterion called demographic parity.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源