论文标题
将句子嵌入到基于LSTM的自回归语言模型中
Fusing Sentence Embeddings Into LSTM-based Autoregressive Language Models
论文作者
论文摘要
尽管蒙面的语言模型具有高度性能,并且被NLP从业人员广泛采用,但它们不能轻易用于自回归语言建模(下一个单词预测和序列概率估计)。我们提出了一种基于LSTM的自回归语言模型,该模型使用融合(例如串联)使用前缀嵌入(来自验证的蒙版语言模型),以获得语言建模的更丰富的上下文表示。我们发现Fusion有助于可靠地降低困惑(16.74 $ \ rightarrow $ 15.80),甚至在从与培训数据的不同域中传输到数据集后,它甚至可以保留。我们还通过将其下一个单词的惊人估计与人类阅读时间相关联,评估了表现最佳的融合模型。与我们的期望相矛盾,尽管整体上的困惑程度有所改善,但相关性仍然与基线模型相同。最后,尽管我们专注于在文本上预先培训的语言模型作为融合的来源,但我们的方法可能会扩展到将表示为固定尺寸矢量表示的任何信息融合到自动回归语言模型中。这些包括例如句子外部信息是为知识库或多模式编码器的表示形式检索的。
Although masked language models are highly performant and widely adopted by NLP practitioners, they can not be easily used for autoregressive language modelling (next word prediction and sequence probability estimation). We present an LSTM-based autoregressive language model which uses prefix embeddings (from a pretrained masked language model) via fusion (e.g. concatenation) to obtain a richer context representation for language modelling. We find that fusion helps reliably in lowering the perplexity (16.74 $\rightarrow$ 15.80), which is even preserved after a transfer to a dataset from a different domain than the training data. We also evaluate the best-performing fusion model by correlating its next word surprisal estimates with human reading times. Contradicting our expectation, and despite the improvement in perplexity overall, the correlation remains the same as for the baseline model. Lastly, while we focus on language models pre-trained on text as the sources for the fusion, our approach can be possibly extended to fuse any information represented as a fixed-size vector into an auto-regressive language model. These include e.g. sentence external information retrieved for a knowledge base or representations of multi-modal encoders.