论文标题
语法信息的问题与异质图变压器回答
Syntax-informed Question Answering with Heterogeneous Graph Transformer
论文作者
论文摘要
大型神经语言模型正在稳步贡献最先进的表现,以应对答案以及其他自然语言和信息处理任务。这些型号的训练价格昂贵。我们建议评估此类预训练的模型是否可以从添加显式语言信息信息中受益,而无需从头开始重新培训。 我们提出了一种语言学知识的答案方法,该方法扩展了基于预训练的变压器的神经语言模型,并使用具有异质图变压器编码的象征性知识。我们通过以依赖关系和组成性图形结构的形式添加句法信息来说明方法。 与伯特(Bert)作为基准的比较经验评估,并以斯坦福大学的问题回答数据集证明了拟议方法的竞争力。我们认为,根据初步实验的进一步结果,我们认为该方法可以扩展到包括语义和语法学在内的进一步语言学信息。
Large neural language models are steadily contributing state-of-the-art performance to question answering and other natural language and information processing tasks. These models are expensive to train. We propose to evaluate whether such pre-trained models can benefit from the addition of explicit linguistics information without requiring retraining from scratch. We present a linguistics-informed question answering approach that extends and fine-tunes a pre-trained transformer-based neural language model with symbolic knowledge encoded with a heterogeneous graph transformer. We illustrate the approach by the addition of syntactic information in the form of dependency and constituency graphic structures connecting tokens and virtual vertices. A comparative empirical performance evaluation with BERT as its baseline and with Stanford Question Answering Dataset demonstrates the competitiveness of the proposed approach. We argue, in conclusion and in the light of further results of preliminary experiments, that the approach is extensible to further linguistics information including semantics and pragmatics.