论文标题
通过视觉和颗粒状方法进行的可解释的机器学习调查超出准解释
Survey of explainable machine learning with visual and granular methods beyond quasi-explanations
论文作者
论文摘要
本文调查了机器学习(ML)解释性的视觉方法(ML),重点是从在ML中占主导地位的准解释转变为由粒状视觉效果支持的特定区域解释。 ML解释从根本上是人类活动,视觉方法更容易解释。尽管存在高维数据的有效视觉表示,但可解释的信息,遮挡和混乱的丢失仍然是一个挑战,这导致了准解释。我们从动机和解释性的不同定义开始。本文着重于准解释和特定域的特定解释之间的明确区别,以及可解释的与实际解释的ML模型之间至关重要的ML模型,这些模型对解释性域至关重要。我们讨论了可解释性,概述视觉解释性的基础,并提出了几种可视化ML模型的方法。接下来,我们介绍了基于最近引入的通用线坐标概念(GLC)的可解释模型的视觉发现方法,重点是可解释的模型。这些方法采取了创建视觉解释的关键步骤,这些解释不仅是准解释,而且是域特定的视觉解释,而这些方法本身是域 - 不可思议的。本文包括基于约翰逊·林登斯特劳斯引理,点对点和点对上的GLC方法以及现实世界中的案例研究的理论限制,以在较低的维度中保留N-D距离。该论文还涵盖了了解ML模型的传统视觉方法,其中包括深度学习和时间序列模型。我们表明,其中许多方法都是准解释,需要进一步的增强才能成为特定领域的解释。我们最后概述了开放问题和当前的研究前沿。
This paper surveys visual methods of explainability of Machine Learning (ML) with focus on moving from quasi-explanations that dominate in ML to domain-specific explanation supported by granular visuals. ML interpretation is fundamentally a human activity and visual methods are more readily interpretable. While efficient visual representations of high-dimensional data exist, the loss of interpretable information, occlusion, and clutter continue to be a challenge, which lead to quasi-explanations. We start with the motivation and the different definitions of explainability. The paper focuses on a clear distinction between quasi-explanations and domain specific explanations, and between explainable and an actually explained ML model that are critically important for the explainability domain. We discuss foundations of interpretability, overview visual interpretability and present several types of methods to visualize the ML models. Next, we present methods of visual discovery of ML models, with the focus on interpretable models, based on the recently introduced concept of General Line Coordinates (GLC). These methods take the critical step of creating visual explanations that are not merely quasi-explanations but are also domain specific visual explanations while these methods themselves are domain-agnostic. The paper includes results on theoretical limits to preserve n-D distances in lower dimensions, based on the Johnson-Lindenstrauss lemma, point-to-point and point-to-graph GLC approaches, and real-world case studies. The paper also covers traditional visual methods for understanding ML models, which include deep learning and time series models. We show that many of these methods are quasi-explanations and need further enhancement to become domain specific explanations. We conclude with outlining open problems and current research frontiers.