论文标题
globenc:通过将整个编码器层纳入变压器中来量化全局令牌归因
GlobEnc: Quantifying Global Token Attribution by Incorporating the Whole Encoder Layer in Transformers
论文作者
论文摘要
对解释变压器的基本动力学的兴趣越来越大。尽管最初认为自我发场模式是主要选择,但最近的研究表明,整合其他组件可以产生更准确的解释。本文介绍了一种新颖的令牌归因分析方法,该方法将所有组件纳入编码器块中,并在整个层中汇总。通过广泛的定量和定性实验,我们证明我们的方法可以产生忠实而有意义的全球令牌归因。我们的实验表明,将几乎每个编码器组件纳入局部(单层)和全局(整个模型)设置都越来越准确地分析。我们的全球归因分析显着优于关于与基于梯度的显着性分数相关的各种任务的先前方法。我们的代码可在https://github.com/mohsenfayyaz/globenc中免费获得。
There has been a growing interest in interpreting the underlying dynamics of Transformers. While self-attention patterns were initially deemed as the primary option, recent studies have shown that integrating other components can yield more accurate explanations. This paper introduces a novel token attribution analysis method that incorporates all the components in the encoder block and aggregates this throughout layers. Through extensive quantitative and qualitative experiments, we demonstrate that our method can produce faithful and meaningful global token attributions. Our experiments reveal that incorporating almost every encoder component results in increasingly more accurate analysis in both local (single layer) and global (the whole model) settings. Our global attribution analysis significantly outperforms previous methods on various tasks regarding correlation with gradient-based saliency scores. Our code is freely available at https://github.com/mohsenfayyaz/GlobEnc.