论文标题
验证的文本排名的变压器:伯特和超越
Pretrained Transformers for Text Ranking: BERT and Beyond
论文作者
论文摘要
文本排名的目的是生成从语料库检索到的查询的有序列表。尽管文本排名最常见的表述是搜索,但在许多自然语言处理应用程序中也可以找到该任务的实例。这项调查提供了文本排名的概述,其神经网络体系结构称为变形金刚,其中伯特是最著名的示例。变形金刚和自我监督预处理的结合是导致自然语言处理(NLP),信息检索(IR)及以后的范式转变的原因。在这项调查中,我们将现有工作的综合作为单一的入学点,为希望更好地理解如何将变形金刚应用于文本排名问题以及希望在这一领域从事工作的研究人员的单一入学点。我们涵盖了广泛的现代技术,分为两个高级类别:在多阶段体系结构中执行重新骑行的变压器模型和直接执行排名的密集检索技术。有两个主题遍布我们的调查:用于处理长文档的技术,超越NLP的典型句子处理,以及解决有效性(即结果质量)和效率(例如查询潜伏期,模型,模型和指数尺寸)之间的权衡方面的技术。尽管变压器体系结构和预训练的技术是最近的创新,但它们如何应用于文本排名的许多方面相对良好地理解并代表成熟的技术。但是,还有许多开放的研究问题,因此,除了为文本排名审计的变压器的基础外,该调查还试图预测该领域的发展方向。
The goal of text ranking is to generate an ordered list of texts retrieved from a corpus in response to a query. Although the most common formulation of text ranking is search, instances of the task can also be found in many natural language processing applications. This survey provides an overview of text ranking with neural network architectures known as transformers, of which BERT is the best-known example. The combination of transformers and self-supervised pretraining has been responsible for a paradigm shift in natural language processing (NLP), information retrieval (IR), and beyond. In this survey, we provide a synthesis of existing work as a single point of entry for practitioners who wish to gain a better understanding of how to apply transformers to text ranking problems and researchers who wish to pursue work in this area. We cover a wide range of modern techniques, grouped into two high-level categories: transformer models that perform reranking in multi-stage architectures and dense retrieval techniques that perform ranking directly. There are two themes that pervade our survey: techniques for handling long documents, beyond typical sentence-by-sentence processing in NLP, and techniques for addressing the tradeoff between effectiveness (i.e., result quality) and efficiency (e.g., query latency, model and index size). Although transformer architectures and pretraining techniques are recent innovations, many aspects of how they are applied to text ranking are relatively well understood and represent mature techniques. However, there remain many open research questions, and thus in addition to laying out the foundations of pretrained transformers for text ranking, this survey also attempts to prognosticate where the field is heading.