论文标题
深度神经网络的信心学习
Confidence-Aware Learning for Deep Neural Networks
论文作者
论文摘要
尽管深层神经网络对各种任务具有力量,但过度强调的预测问题限制了它们在许多安全至关重要的应用中的实际使用。已经提出了许多最近的工作来减轻此问题,但是其中大多数要求培训和/或推理阶段或定制体系结构的额外计算成本,以分别输出置信度估计。在本文中,我们提出了一种训练具有新颖损失函数的深神经网络的方法,称为正确性排名损失,该方法明确规定了班级概率,以根据置信度的序数排名来估计更好的置信度估计。提出的方法易于实现,可以应用于现有体系结构而无需任何修改。此外,它的培训计算成本几乎与传统的深层分类器相同,并且通过单个推论进行可靠的预测。分类基准数据集的广泛实验结果表明,该建议的方法有助于网络产生良好的置信度估计。我们还证明,这对于与置信度估计,分布外检测和积极学习密切相关的任务有效。
Despite the power of deep neural networks for a wide range of tasks, an overconfident prediction issue has limited their practical use in many safety-critical applications. Many recent works have been proposed to mitigate this issue, but most of them require either additional computational costs in training and/or inference phases or customized architectures to output confidence estimates separately. In this paper, we propose a method of training deep neural networks with a novel loss function, named Correctness Ranking Loss, which regularizes class probabilities explicitly to be better confidence estimates in terms of ordinal ranking according to confidence. The proposed method is easy to implement and can be applied to the existing architectures without any modification. Also, it has almost the same computational costs for training as conventional deep classifiers and outputs reliable predictions by a single inference. Extensive experimental results on classification benchmark datasets indicate that the proposed method helps networks to produce well-ranked confidence estimates. We also demonstrate that it is effective for the tasks closely related to confidence estimation, out-of-distribution detection and active learning.