论文标题
使用非负概念激活向量和CNN模型的决策树的基于概念的解释
Concept-based Explanations using Non-negative Concept Activation Vectors and Decision Tree for CNN Models
论文作者
论文摘要
本文评估了基于从基于概念的解释器中提取的概念训练决策树是否可以提高卷积神经网络(CNNS)模型的解释性,并提高使用解释器的忠诚度和性能。计算机视觉的CNN在关键行业中表现出色。但是,由于其复杂性和缺乏解释性,在部署CNN时,这是一个重大障碍。解释计算机视觉模型的最新研究已从提取低级特征(基于像素的解释)转变为中或高级特征(基于概念的解释)。当前的研究方向倾向于在开发近似算法(例如线性或决策树模型)中使用提取的特征来解释原始模型。在这项工作中,我们修改了基于概念的最先进的解释之一,并提出了一个名为Treeice的替代框架。我们根据忠诚度(与原始模型的标签的近似模型),性能(与地面真相标签的近似模型)和可解释性(对人类的近似模型有意义)设计了系统评估。我们进行计算评估(对于忠诚度和表现)和人类主题实验(对于解释性),我们发现树冰的表现优于可解释性的基线,并以语义树结构的形式产生了更多人类可读的解释。这项工作的特点是,当解释性至关重要时,拥有更易于理解的解释非常重要。
This paper evaluates whether training a decision tree based on concepts extracted from a concept-based explainer can increase interpretability for Convolutional Neural Networks (CNNs) models and boost the fidelity and performance of the used explainer. CNNs for computer vision have shown exceptional performance in critical industries. However, it is a significant barrier when deploying CNNs due to their complexity and lack of interpretability. Recent studies to explain computer vision models have shifted from extracting low-level features (pixel-based explanations) to mid-or high-level features (concept-based explanations). The current research direction tends to use extracted features in developing approximation algorithms such as linear or decision tree models to interpret an original model. In this work, we modify one of the state-of-the-art concept-based explanations and propose an alternative framework named TreeICE. We design a systematic evaluation based on the requirements of fidelity (approximate models to original model's labels), performance (approximate models to ground-truth labels), and interpretability (meaningful of approximate models to humans). We conduct computational evaluation (for fidelity and performance) and human subject experiments (for interpretability) We find that Tree-ICE outperforms the baseline in interpretability and generates more human readable explanations in the form of a semantic tree structure. This work features how important to have more understandable explanations when interpretability is crucial.