论文标题
通过转移学习的情感意识到的多任务处理方法,用于虚假新闻和谣言检测
An Emotion-Aware Multi-Task Approach to Fake News and Rumour Detection using Transfer Learning
论文作者
论文摘要
社交网站,博客和在线文章是全球互联网用户的即时新闻来源。但是,在没有严格的法规规定社交媒体上每个文本的真实性的情况下,其中一些文本很可能是假新闻或谣言。他们的欺骗性和立即传播的能力会对社会产生不利影响。这需要需要更有效地发现虚假新闻和网络上的谣言。在这项工作中,我们使用转移学习来注释四个假新闻检测和谣言检测数据集。我们展示了文本的合法性与其在假新闻和谣言检测中的内在情感之间的相关性,并证明即使在同一情感类别中,假新闻和真实新闻通常也会有所不同,可以用来改善功能提取。基于此,我们为假新闻和谣言检测提出了一个多任务框架,可以预测文本的情感和合法性。我们在单个任务和多任务设置中培训各种深度学习模型,以进行更全面的比较。我们进一步分析了我们在跨域设置中进行虚假新闻检测的多任务方法的性能,以验证其在跨数据集中更好地概括的功效,并验证情绪充当域独立的功能。实验结果验证了我们的多任务模型在准确性,精度,召回和F1分数方面始终优于其单任务对应,无论是在域内和跨域设置而言。我们还定性地分析了单任务和多任务学习模型的性能差异。
Social networking sites, blogs, and online articles are instant sources of news for internet users globally. However, in the absence of strict regulations mandating the genuineness of every text on social media, it is probable that some of these texts are fake news or rumours. Their deceptive nature and ability to propagate instantly can have an adverse effect on society. This necessitates the need for more effective detection of fake news and rumours on the web. In this work, we annotate four fake news detection and rumour detection datasets with their emotion class labels using transfer learning. We show the correlation between the legitimacy of a text with its intrinsic emotion for fake news and rumour detection, and prove that even within the same emotion class, fake and real news are often represented differently, which can be used for improved feature extraction. Based on this, we propose a multi-task framework for fake news and rumour detection, predicting both the emotion and legitimacy of the text. We train a variety of deep learning models in single-task and multi-task settings for a more comprehensive comparison. We further analyze the performance of our multi-task approach for fake news detection in cross-domain settings to verify its efficacy for better generalization across datasets, and to verify that emotions act as a domain-independent feature. Experimental results verify that our multi-task models consistently outperform their single-task counterparts in terms of accuracy, precision, recall, and F1 score, both for in-domain and cross-domain settings. We also qualitatively analyze the difference in performance in single-task and multi-task learning models.