论文标题
分解分布检测:许多基于OOD培训数据的方法估计相同数量的组合
Breaking Down Out-of-Distribution Detection: Many Methods Based on OOD Training Data Estimate a Combination of the Same Core Quantities
论文作者
论文摘要
在值得信赖的机器学习中,这是一个重要的问题,可以识别与分布任务无关的输入的分布(OOD)输入。近年来,已经提出了许多分布式检测方法。本文的目的是确定共同的目标以及确定不同OOD检测方法的隐式评分函数。我们专注于在培训期间使用替代OOD数据的方法的子类,以学习在测试时概括为新看不见的远离分布的OOD检测分数。我们表明,内部和(不同)外部分布之间的二进制歧视等同于OOD检测问题的几种不同的表述。当与标准分类器以共同的方式接受培训时,该二进制判别器达到了与离群暴露相似的OOD检测性能。此外,我们表明,在训练和测试外部分布相同的情况下,异常敞口使用的置信度损失与理论上最佳的评分功能不同,这与理论上最佳的评分函数不同,这又与训练基于能量的OOD检测器或添加类别时使用的情况相似。在实践中,如果以完全相同的方式培训,所有这些方法的性能类似。
It is an important problem in trustworthy machine learning to recognize out-of-distribution (OOD) inputs which are inputs unrelated to the in-distribution task. Many out-of-distribution detection methods have been suggested in recent years. The goal of this paper is to recognize common objectives as well as to identify the implicit scoring functions of different OOD detection methods. We focus on the sub-class of methods that use surrogate OOD data during training in order to learn an OOD detection score that generalizes to new unseen out-distributions at test time. We show that binary discrimination between in- and (different) out-distributions is equivalent to several distinct formulations of the OOD detection problem. When trained in a shared fashion with a standard classifier, this binary discriminator reaches an OOD detection performance similar to that of Outlier Exposure. Moreover, we show that the confidence loss which is used by Outlier Exposure has an implicit scoring function which differs in a non-trivial fashion from the theoretically optimal scoring function in the case where training and test out-distribution are the same, which again is similar to the one used when training an Energy-Based OOD detector or when adding a background class. In practice, when trained in exactly the same way, all these methods perform similarly.