论文标题
Tsinsight:时间序列数据中可解释性的本地全球归因框架
TSInsight: A local-global attribution framework for interpretability in time-series data
论文作者
论文摘要
随着在安全至关重要方案中使用深度学习方法的使用兴起,可解释性比以往任何时候都更为重要。尽管已经探索了有关视觉方式的许多不同方向,但是由于其可理解性差而仅通过少数测试的方法忽略了时间序列数据。我们通过提出tsinsight以新颖的方式解决可解释性的问题,在该问题上我们将自动编码器附加到分类器上,并通过稀疏性诱导其输出范围,并根据分类器的梯度和重建惩罚来微调IT。 Tsinsight学会了保留对分类器预测至关重要的特征,并抑制那些无关紧要的功能,即作为提高可解释性的特征归因方法。与大多数其他归因框架相反,Tsinsight能够同时生成基于实例的和基于模型的解释。我们在8个不同的时间序列数据集上评估了Tsinsight以及其他9种常用的归因方法,以验证其疗效。评估结果表明,Tsinsight自然实现了输出空间收缩,这是用于可解释深度序列模型的有效工具。
With the rise in the employment of deep learning methods in safety-critical scenarios, interpretability is more essential than ever before. Although many different directions regarding interpretability have been explored for visual modalities, time-series data has been neglected with only a handful of methods tested due to their poor intelligibility. We approach the problem of interpretability in a novel way by proposing TSInsight where we attach an auto-encoder to the classifier with a sparsity-inducing norm on its output and fine-tune it based on the gradients from the classifier and a reconstruction penalty. TSInsight learns to preserve features that are important for prediction by the classifier and suppresses those that are irrelevant i.e. serves as a feature attribution method to boost interpretability. In contrast to most other attribution frameworks, TSInsight is capable of generating both instance-based and model-based explanations. We evaluated TSInsight along with 9 other commonly used attribution methods on 8 different time-series datasets to validate its efficacy. Evaluation results show that TSInsight naturally achieves output space contraction, therefore, is an effective tool for the interpretability of deep time-series models.