论文标题

更少学习快捷方式:分析和减轻伪造特征标签相关的学习

Less Learn Shortcut: Analyzing and Mitigating Learning of Spurious Feature-Label Correlation

论文作者

Du, Yanrui, Yan, Jing, Chen, Yan, Liu, Jing, Zhao, Sendong, She, Qiaoqiao, Wu, Hua, Wang, Haifeng, Qin, Bing

论文摘要

最近的研究表明,深层神经网络通常将数据集偏见作为做出决定而不是理解任务的捷径,从而导致现实应用程序中的失败。在这项研究中,我们专注于单词特征与标签之间的虚假相关性,这些标签与培训数据的偏见数据分布中学习。特别是,我们将单词与特定的标签高度共同相处定义为有偏见的单词,并将包含有偏见的单词作为有偏见的示例定义。我们的分析表明,模型更容易学习,而在预测时,有偏见的单词对模型的预测做出了更高的贡献,并且模型倾向于将预测的标签分配在单词和标签之间的虚假相关性上。为了减轻模型对快捷方式的过度依赖(即虚假相关性),我们提出了一种训练策略较少的毛线效应(LLS):我们的策略量化了有偏见的例子的有偏见程度,并对它们进行了相应的重视。有关问题匹配,自然推理和情感分析任务的实验结果表明,LLS是一种任务不合时宜的策略,可以在对抗数据上提高模型性能,同时保持对内域数据的良好性能。

Recent research has revealed that deep neural networks often take dataset biases as a shortcut to make decisions rather than understand tasks, leading to failures in real-world applications. In this study, we focus on the spurious correlation between word features and labels that models learn from the biased data distribution of training data. In particular, we define the word highly co-occurring with a specific label as biased word, and the example containing biased word as biased example. Our analysis shows that biased examples are easier for models to learn, while at the time of prediction, biased words make a significantly higher contribution to the models' predictions, and models tend to assign predicted labels over-relying on the spurious correlation between words and labels. To mitigate models' over-reliance on the shortcut (i.e. spurious correlation), we propose a training strategy Less-Learn-Shortcut (LLS): our strategy quantifies the biased degree of the biased examples and down-weights them accordingly. Experimental results on Question Matching, Natural Language Inference and Sentiment Analysis tasks show that LLS is a task-agnostic strategy and can improve the model performance on adversarial data while maintaining good performance on in-domain data.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源