论文标题
适应性融合:转移学习的非破坏性任务组成
AdapterFusion: Non-Destructive Task Composition for Transfer Learning
论文作者
论文摘要
顺序的微调和多任务学习是旨在纳入多个任务知识的方法。但是,他们遭受了灾难性的遗忘和数据集平衡困难。为了解决这些缺点,我们提出了AdapterFusion,这是一种新的两个阶段学习算法,从多个任务中利用知识。首先,在知识提取阶段,我们学习任务特定的参数,称为适配器,该参数封装了特定于任务的信息。然后,我们将适配器组合在单独的知识组成步骤中。我们表明,通过将两个阶段分开,即知识提取和知识组成,分类器可以有效利用以非损害方式从多个任务中学到的表示形式。我们从经验上评估了16个不同NLU任务的适应性融合,并发现它有效地结合了模型不同层的各种知识。我们表明,我们的方法优于传统策略,例如完整的微调以及多任务学习。我们的代码和适配器可在adapterhub.ml上找到。
Sequential fine-tuning and multi-task learning are methods aiming to incorporate knowledge from multiple tasks; however, they suffer from catastrophic forgetting and difficulties in dataset balancing. To address these shortcomings, we propose AdapterFusion, a new two stage learning algorithm that leverages knowledge from multiple tasks. First, in the knowledge extraction stage we learn task specific parameters called adapters, that encapsulate the task-specific information. We then combine the adapters in a separate knowledge composition step. We show that by separating the two stages, i.e., knowledge extraction and knowledge composition, the classifier can effectively exploit the representations learned from multiple tasks in a non-destructive manner. We empirically evaluate AdapterFusion on 16 diverse NLU tasks, and find that it effectively combines various types of knowledge at different layers of the model. We show that our approach outperforms traditional strategies such as full fine-tuning as well as multi-task learning. Our code and adapters are available at AdapterHub.ml.