论文标题

切片概率差异的统计和拓扑特性

Statistical and Topological Properties of Sliced Probability Divergences

论文作者

Nadjahi, Kimia, Durmus, Alain, Chizat, Lénaïc, Kolouri, Soheil, Shahrampour, Shahin, Şimşekli, Umut

论文摘要

在比较各种机器学习应用程序中的两个概率度量(包括生成建模)中,切片差异的想法已被证明是成功的,并在于计算这两个度量的一维随机投影之间的“基础差异”的期望值。但是,该技术的拓扑,统计和计算后果尚未得到良好的建立。在本文中,我们旨在弥合这一差距,并得出切成概率差异的各种理论特性。首先,我们表明切片保留了公理和差异的弱连续性,这意味着切成薄片的差异将具有相似的拓扑特性。然后,在基本差异属于积分概率指标类别的情况下,我们将结果确定。另一方面,我们确定在轻度条件下,切成差的样本复杂性不取决于问题维度。我们最终将一般结果应用于几个基本差异,并说明了我们关于合成和真实数据实验的理论。

The idea of slicing divergences has been proven to be successful when comparing two probability measures in various machine learning applications including generative modeling, and consists in computing the expected value of a `base divergence' between one-dimensional random projections of the two measures. However, the topological, statistical, and computational consequences of this technique have not yet been well-established. In this paper, we aim at bridging this gap and derive various theoretical properties of sliced probability divergences. First, we show that slicing preserves the metric axioms and the weak continuity of the divergence, implying that the sliced divergence will share similar topological properties. We then precise the results in the case where the base divergence belongs to the class of integral probability metrics. On the other hand, we establish that, under mild conditions, the sample complexity of a sliced divergence does not depend on the problem dimension. We finally apply our general results to several base divergences, and illustrate our theory on both synthetic and real data experiments.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源