论文标题
停止过滤:多视图属性增强对话学习
Stop Filtering: Multi-View Attribute-Enhanced Dialogue Learning
论文作者
论文摘要
通过过滤原始的对话语料库来提高模型的对话能力,人们越来越兴趣。以前的过滤策略通常依靠评分方法来评估和丢弃样本,从而使模型更容易增强相应的对话属性(例如一致性)。但是,废弃的样本可能会从其他角度获得高分,并可以对模型学习提供正则化影响,这会导致性能改善对滤波比敏感。在这项工作中,我们提出了一个多视图属性增强的对话学习框架,该框架可以更加和全面地增强与属性相关的功能。我们的框架没有过滤原始数据集来训练模型,而是首先预先培训原始数据集上的模型,然后通过所选子集的适配器进行微调,这也增强了响应的某些属性,但没有上述问题。考虑到对话属性的多样性,我们进一步设计了一种多视图增强机制,包括多视图选择和跨视图融合。 IT分别从多个角度将高质量样本分组,并通过相应的样本集和适配器增强了响应的不同属性,使知识独立并允许灵活地集成。经验结果和分析表明,我们的框架可以在增强对话属性和融合特定视图的知识方面显着提高性能。
There is a growing interest in improving the conversational ability of models by filtering the raw dialogue corpora. Previous filtering strategies usually rely on a scoring method to assess and discard samples from one perspective, enabling the model to enhance the corresponding dialogue attributes (e.g., consistency) more easily. However, the discarded samples may obtain high scores in other perspectives and can provide regularization effects on the model learning, which causes the performance improvement to be sensitive to the filtering ratio. In this work, we propose a multi-view attribute-enhanced dialogue learning framework that strengthens the attribute-related features more robustly and comprehensively. Instead of filtering the raw dataset to train the model, our framework first pre-trains the model on the raw dataset and then fine-tunes it through adapters on the selected sub-sets, which also enhances certain attributes of responses but without suffering from the problems mentioned above. Considering the variety of the dialogue attribute, we further design a multi-view enhancement mechanism, including multi-view selection and inter-view fusion. It groups the high-quality samples from multiple perspectives, respectively, and enhances different attributes of responses with the corresponding sample sets and adapters, keeping knowledge independent and allowing flexible integration. Empirical results and analysis show that our framework can improve the performance significantly in terms of enhancing dialogue attributes and fusing view-specific knowledge.