论文标题

为什么要模型?评估石灰的优势和局限性

Why model why? Assessing the strengths and limitations of LIME

论文作者

Dieber, Jürgen, Kirrane, Sabrina

论文摘要

当涉及复杂的机器学习模型(通常称为黑匣子)时,了解基本决策过程对于诸如医疗保健和金融服务等领域以及与安全关键系统(例如自动驾驶汽车)相关的领域至关重要。近年来,人们对可解释的人工智能(XAI)工具和技术的兴趣有所增加。但是,现有的XAI框架的有效性,尤其是与与图像相比的数据合作的算法,仍然是一个开放的研究问题。为了解决这一差距,在本文中,我们研究了局部可解释的模型不可解释(Lime)XAI框架的有效性,XAI框架是文献中最受欢迎的Model Model docnostic框架之一,在使表格模型更加可解释的方面,其特定侧重于其性能。特别是,我们在表格数据集上应用了几种最先进的机器学习算法,并演示如何使用石灰来补充常规性能评估方法。此外,我们通过可用性研究评估了石灰产生的产出的可理解性,涉及不熟悉石灰的参与者,以及通过评估框架的总体可用性,该框架源自国际标准化组织9241-11-11:1998标准。

When it comes to complex machine learning models, commonly referred to as black boxes, understanding the underlying decision making process is crucial for domains such as healthcare and financial services, and also when it is used in connection with safety critical systems such as autonomous vehicles. As such interest in explainable artificial intelligence (xAI) tools and techniques has increased in recent years. However, the effectiveness of existing xAI frameworks, especially concerning algorithms that work with data as opposed to images, is still an open research question. In order to address this gap, in this paper we examine the effectiveness of the Local Interpretable Model-Agnostic Explanations (LIME) xAI framework, one of the most popular model agnostic frameworks found in the literature, with a specific focus on its performance in terms of making tabular models more interpretable. In particular, we apply several state of the art machine learning algorithms on a tabular dataset, and demonstrate how LIME can be used to supplement conventional performance assessment methods. In addition, we evaluate the understandability of the output produced by LIME both via a usability study, involving participants who are not familiar with LIME, and its overall usability via an assessment framework, which is derived from the International Organisation for Standardisation 9241-11:1998 standard.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源