论文标题

策略研究:学习掩盖闭幕质量检查

Studying Strategically: Learning to Mask for Closed-book QA

论文作者

Ye, Qinyuan, Li, Belinda Z., Wang, Sinong, Bolte, Benjamin, Ma, Hao, Yih, Wen-tau, Ren, Xiang, Khabsa, Madian

论文摘要

封闭式问题撤职(QA)是一项艰巨的任务,需要模型直接回答问题而无需访问外部知识。已经表明,通过(问题,答案)示例直接微调预训练的语言模型会产生令人惊讶的竞争性能,通过在一般预训练和微调之间添加中间的预训练阶段,可以进一步改善。先前的工作在此中间阶段使用了启发式,命名实体和日期被掩盖,并培训了该模型以恢复这些令牌。在本文中,我们旨在学习中级预训练阶段的最佳掩蔽策略。我们首先使用下游任务本身的监督,然后在中级预训练期间部署学习的策略,从而提取可能经过测试的跨度的跨度策略。因此,我们的策略将与任务相关的知识包装到语言模型的参数中。我们的方法对Triviaqa尤其有效,在用于预训练BART的情况下表现优于强大的启发式方法。

Closed-book question-answering (QA) is a challenging task that requires a model to directly answer questions without access to external knowledge. It has been shown that directly fine-tuning pre-trained language models with (question, answer) examples yields surprisingly competitive performance, which is further improved upon through adding an intermediate pre-training stage between general pre-training and fine-tuning. Prior work used a heuristic during this intermediate stage, whereby named entities and dates are masked, and the model is trained to recover these tokens. In this paper, we aim to learn the optimal masking strategy for the intermediate pre-training stage. We first train our masking policy to extract spans that are likely to be tested, using supervision from the downstream task itself, then deploy the learned policy during intermediate pre-training. Thus, our policy packs task-relevant knowledge into the parameters of a language model. Our approach is particularly effective on TriviaQA, outperforming strong heuristics when used to pre-train BART.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源