论文标题
事件参数提取的文本需要:多源学习的零和几次射击
Textual Entailment for Event Argument Extraction: Zero- and Few-Shot with Multi-Source Learning
论文作者
论文摘要
最近的工作表明,诸如关系提取(RE)之类的NLP任务可以使用语言来重塑为文本需要任务,并且由于预先训练的需要模型,在零拍摄和少量设置方面的性能很强。当前RE数据集中的关系很容易被口头表演的事实,这一事实对是否有效在更复杂的任务中有效。在这项工作中,我们表明,在事件论点提取(EAE)中,元素在ACE和Wikievents中分别将手动注释的需求分别减少到50%和20%,同时与全面培训相同的表现。更重要的是,我们表明将EAE重新铸造,因为需要减轻对图案的依赖,这是用于在域之间转移注释的路障。得益于这一目标,ACE和Wikivents之间的多源转移将注释进一步降低至全部培训的10%和5%,而无需转移。我们的分析表明,良好结果的关键是使用多个元素数据集预先培训元素模型。与以前的方法相似,我们的方法需要少量的手动语言努力:每个事件参数类型只需少于15分钟,并且可以与具有不同专业知识水平的用户获得可比的结果。
Recent work has shown that NLP tasks such as Relation Extraction (RE) can be recasted as Textual Entailment tasks using verbalizations, with strong performance in zero-shot and few-shot settings thanks to pre-trained entailment models. The fact that relations in current RE datasets are easily verbalized casts doubts on whether entailment would be effective in more complex tasks. In this work we show that entailment is also effective in Event Argument Extraction (EAE), reducing the need of manual annotation to 50% and 20% in ACE and WikiEvents respectively, while achieving the same performance as with full training. More importantly, we show that recasting EAE as entailment alleviates the dependency on schemas, which has been a road-block for transferring annotations between domains. Thanks to the entailment, the multi-source transfer between ACE and WikiEvents further reduces annotation down to 10% and 5% (respectively) of the full training without transfer. Our analysis shows that the key to good results is the use of several entailment datasets to pre-train the entailment model. Similar to previous approaches, our method requires a small amount of effort for manual verbalization: only less than 15 minutes per event argument type is needed, and comparable results can be achieved with users with different level of expertise.