论文标题
带有顺序跨模式语义图的目标情感分类
Target-oriented Sentiment Classification with Sequential Cross-modal Semantic Graph
论文作者
论文摘要
基于多模式方面的情感分类(MABSC)是对句子和图像中提到的目标实体的情感进行分类的任务。但是,以前的方法未能说明图像和文本之间的细粒语义关联,从而导致对细粒图像方面和观点的识别有限。为了解决这些局限性,在本文中,我们提出了一种称为SEQCSG的新方法,该方法使用顺序跨模式语义图增强了编码器 - 码头模糊分类框架。 SEQCSG利用图像标题和场景图来提取全局和本地细粒图像信息,并将它们视为跨模式语义图的元素以及来自Tweets的令牌。顺序跨模式语义图表示为一个序列,该序列具有多模式的邻接矩阵,指示元素之间的关系。实验结果表明,该方法的表现优于现有方法,并在两个标准数据集上实现最先进的性能。进一步的分析表明,该模型可以隐式地学习图像的细粒信息与给定目标的文本之间的相关性。我们的代码可在https://github.com/zjukg/seqcsg上找到。
Multi-modal aspect-based sentiment classification (MABSC) is task of classifying the sentiment of a target entity mentioned in a sentence and an image. However, previous methods failed to account for the fine-grained semantic association between the image and the text, which resulted in limited identification of fine-grained image aspects and opinions. To address these limitations, in this paper we propose a new approach called SeqCSG, which enhances the encoder-decoder sentiment classification framework using sequential cross-modal semantic graphs. SeqCSG utilizes image captions and scene graphs to extract both global and local fine-grained image information and considers them as elements of the cross-modal semantic graph along with tokens from tweets. The sequential cross-modal semantic graph is represented as a sequence with a multi-modal adjacency matrix indicating relationships between elements. Experimental results show that the approach outperforms existing methods and achieves state-of-the-art performance on two standard datasets. Further analysis has demonstrated that the model can implicitly learn the correlation between fine-grained information of the image and the text with the given target. Our code is available at https://github.com/zjukg/SeqCSG.