论文标题

深度自适应图形循环网络用于文本分类

Depth-Adaptive Graph Recurrent Network for Text Classification

论文作者

Liu, Yijin, Meng, Fandong, Chen, Yufeng, Xu, Jinan, Zhou, Jie

论文摘要

句子状态LSTM(S-LSTM)是一个功能强大且高效的图形循环网络,该网络将单词视为节点,并同时在层次上执行层的复发步骤。尽管在文本表示上取得了成功,但S-LSTM仍然遭受两个缺点。首先,给定句子,某些单词通常比其他单词更模棱两可,因此,对于这些困难单词,需要采取更多的计算步骤,反之亦然。但是,S-LSTM对所有单词都采取固定的计算步骤,而不论其硬度如何。次要信息来自缺乏对自然语言本质上重要的顺序信息(例如单词顺序)。在本文中,我们试图解决这些问题,并为S-LSTM提出一种深度自适应机制,该机制使模型可以根据需要学习对不同单词进行执行的计算步骤。此外,我们将额外的RNN层集成到注入顺序信息,这也是自适应深度决策的输入功能。经典文本分类任务(各种尺寸和域中的24个数据集)的结果表明,我们的模型对传统的S-LSTM和其他高性能模型(例如变压器)带来了重大改进,同时实现了良好的准确速度贸易。

The Sentence-State LSTM (S-LSTM) is a powerful and high efficient graph recurrent network, which views words as nodes and performs layer-wise recurrent steps between them simultaneously. Despite its successes on text representations, the S-LSTM still suffers from two drawbacks. Firstly, given a sentence, certain words are usually more ambiguous than others, and thus more computation steps need to be taken for these difficult words and vice versa. However, the S-LSTM takes fixed computation steps for all words, irrespective of their hardness. The secondary one comes from the lack of sequential information (e.g., word order) that is inherently important for natural language. In this paper, we try to address these issues and propose a depth-adaptive mechanism for the S-LSTM, which allows the model to learn how many computational steps to conduct for different words as required. In addition, we integrate an extra RNN layer to inject sequential information, which also serves as an input feature for the decision of adaptive depths. Results on the classic text classification task (24 datasets in various sizes and domains) show that our model brings significant improvements against the conventional S-LSTM and other high-performance models (e.g., the Transformer), meanwhile achieving a good accuracy-speed trade off.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源