论文标题
不要骗我!通过经过验证的扰动分析,可靠和有效的解释性
Don't Lie to Me! Robust and Efficient Explainability with Verified Perturbation Analysis
论文作者
论文摘要
已经提出了多种方法来试图解释其决策的深层神经网络。这些方法的关键是需要有效地品尝像素空间以得出重要性图。但是,已经表明,用于日期的抽样方法引入了偏见和其他工件,从而导致对单个像素的重要性的估计不准确,并严重限制了当前解释性方法的可靠性。不幸的是,替代方案 - 详尽地采样图像空间在计算上是过度的。在本文中,我们介绍了EVA(使用经过验证的扰动分析解释) - 第一个解释性方法保证对扰动空间进行详尽的探索。具体而言,我们利用经过验证的扰动分析的有益属性 - 时间效率,易处理性和保证对歧管的完全覆盖范围 - 有效地表征了最有可能驱动模型决策的输入变量。我们系统地评估了该方法,并在多个基准测试中证明了最新结果。
A variety of methods have been proposed to try to explain how deep neural networks make their decisions. Key to those approaches is the need to sample the pixel space efficiently in order to derive importance maps. However, it has been shown that the sampling methods used to date introduce biases and other artifacts, leading to inaccurate estimates of the importance of individual pixels and severely limit the reliability of current explainability methods. Unfortunately, the alternative -- to exhaustively sample the image space is computationally prohibitive. In this paper, we introduce EVA (Explaining using Verified perturbation Analysis) -- the first explainability method guarantee to have an exhaustive exploration of a perturbation space. Specifically, we leverage the beneficial properties of verified perturbation analysis -- time efficiency, tractability and guaranteed complete coverage of a manifold -- to efficiently characterize the input variables that are most likely to drive the model decision. We evaluate the approach systematically and demonstrate state-of-the-art results on multiple benchmarks.