论文标题
互动解释模型分析的语法
The Grammar of Interactive Explanatory Model Analysis
论文作者
论文摘要
对预测模型的深入分析的需求日益增长,导致了一系列解释其本地和全球性质的新方法。以下哪种方法是最好的?事实证明,这是一个不适的问题。一个人无法使用仅提供一个视角的单一方法来充分解释黑盒机器学习模型。孤立的解释容易误解,导致错误或简单的推理。这个问题称为Rashomon效应,是指对同一现象的多样化,甚至是矛盾的解释。令人惊讶的是,大多数用于解释和负责任的机器学习开发的方法都集中在对模型行为的单一方面。相比之下,我们将解释性作为模型的互动和顺序分析的问题。本文提出了不同的解释模型分析(EMA)方法如何相互补充,并讨论了为什么将它们并列。引入的交互式EMA(IEMA)的过程源自可解释的机器学习的算法一面,旨在拥抱认知科学中发展的思想。我们将IEMA的语法形式化,以描述潜在的人类模型对话。它是在以人为中心的开源软件框架中实现的,该框架采用交互性,可定制性和自动化作为其主要特征。我们进行了一项用户研究以评估IEMA的实用性,这表明对模型的互动顺序分析增加了人类决策的绩效和信心。
The growing need for in-depth analysis of predictive models leads to a series of new methods for explaining their local and global properties. Which of these methods is the best? It turns out that this is an ill-posed question. One cannot sufficiently explain a black-box machine learning model using a single method that gives only one perspective. Isolated explanations are prone to misunderstanding, leading to wrong or simplistic reasoning. This problem is known as the Rashomon effect and refers to diverse, even contradictory, interpretations of the same phenomenon. Surprisingly, most methods developed for explainable and responsible machine learning focus on a single-aspect of the model behavior. In contrast, we showcase the problem of explainability as an interactive and sequential analysis of a model. This paper proposes how different Explanatory Model Analysis (EMA) methods complement each other and discusses why it is essential to juxtapose them. The introduced process of Interactive EMA (IEMA) derives from the algorithmic side of explainable machine learning and aims to embrace ideas developed in cognitive sciences. We formalize the grammar of IEMA to describe potential human-model dialogues. It is implemented in a widely used human-centered open-source software framework that adopts interactivity, customizability and automation as its main traits. We conduct a user study to evaluate the usefulness of IEMA, which indicates that an interactive sequential analysis of a model increases the performance and confidence of human decision making.