论文标题
意大利面:桌面操作通过句子桌子披肩预训练的意识到事实验证
PASTA: Table-Operations Aware Fact Verification via Sentence-Table Cloze Pre-training
论文作者
论文摘要
事实验证最近引起了很多研究的关注,例如新闻,市场营销和决策,因为在线错误信息和虚假信息可能会影响人们的意见并影响人们的行动。虽然事实检查是一项艰巨的任务,但在许多情况下,可以根据表格上的分析可以轻松地揭穿虚假陈述,并提供可靠的信息。因此,基于桌子的事实验证最近已成为一个重要且不断增长的研究领域。但是,由于缺乏可用于预先训练语言模型(LMS)来了解常见的表操作的数据集,例如汇总列或比较元组,进度受到限制。为了弥合这一差距,在本文中,我们介绍了Pasta,这是一个新颖的最新框架,用于通过综合句子桌子披肩问题进行预训练,用于基于桌子的事实验证。特别是,我们设计了六种类型的常见句子披肩任务,包括过滤器,聚合,最高级,比较,序数和独特的任务,我们基于我们合成一个由wikitables的120万个句子对的大型语料库。意大利面使用最近训练的LM,Debertav3,并在我们的语料库上进一步预处理。我们的实验结果表明,面食在两个基于表的事实验证基准:tabfact和sem-tab fats上实现了新的最新性能。特别是,在包含多个操作的复杂的一组TABFACT上,面食在很大程度上优于先前的最新水平(85.6%vs. 80.9%),而小型TabFact测试集中的面食和人类性能之间的差距则缩小到1.5分(90.6%vs. 92.1%)。
Fact verification has attracted a lot of research attention recently, e.g., in journalism, marketing, and policymaking, as misinformation and disinformation online can sway one's opinion and affect one's actions. While fact-checking is a hard task in general, in many cases, false statements can be easily debunked based on analytics over tables with reliable information. Hence, table-based fact verification has recently emerged as an important and growing research area. Yet, progress has been limited due to the lack of datasets that can be used to pre-train language models (LMs) to be aware of common table operations, such as aggregating a column or comparing tuples. To bridge this gap, in this paper we introduce PASTA, a novel state-of-the-art framework for table-based fact verification via pre-training with synthesized sentence-table cloze questions. In particular, we design six types of common sentence-table cloze tasks, including Filter, Aggregation, Superlative, Comparative, Ordinal, and Unique, based on which we synthesize a large corpus consisting of 1.2 million sentence-table pairs from WikiTables. PASTA uses a recent pre-trained LM, DeBERTaV3, and further pretrains it on our corpus. Our experimental results show that PASTA achieves new state-of-the-art performance on two table-based fact verification benchmarks: TabFact and SEM-TAB-FACTS. In particular, on the complex set of TabFact, which contains multiple operations, PASTA largely outperforms the previous state of the art by 4.7 points (85.6% vs. 80.9%), and the gap between PASTA and human performance on the small TabFact test set is narrowed to just 1.5 points (90.6% vs. 92.1%).