论文标题

参数有效的转移从顺序行为转移用于用户建模和建议

Parameter-Efficient Transfer from Sequential Behaviors for User Modeling and Recommendation

论文作者

Yuan, Fajie, He, Xiangnan, Karatzoglou, Alexandros, Zhang, Liguang

论文摘要

归纳转移学习对计算机视觉和NLP域有很大影响,但在推荐系统领域尚未使用。尽管基于建模用户项目互动序列生成建议的研究已经进行了大量研究,但其中很少有人试图代表和传输这些模型,以服务于仅存在有限数据的下游任务。 在本文中,我们探讨了有效学习可以应用于各种任务的单一用户表示的任务,从跨域建议到用户配置文件预测。微调大型的预训练网络并将其调整为下游任务是解决此类任务的有效方法。但是,考虑到每个新任务都需要重新训练整个模型,微型调整是效率低下的。为了克服此问题,我们开发了一个参数有效的传输学习体系结构,称为PETERREC,可以在各种下游任务中直接配置。具体而言,Peterrec允许在微调过程中通过注入一系列重新学习的神经网络,在微调过程中保持不变,这些神经网络很小,但与学习整个网络一样表现力。我们执行广泛的实验消融,以在五个下游任务中显示出学习的用户表示的有效性。此外,我们表明Peterrec在多个域中执行有效的传输学习,在这些域中,与对整个模型参数相比,它具有可比性或有时更好的性能。代码和数据集可在https://github.com/fajieyuan/sigir2020_peterrec上找到。

Inductive transfer learning has had a big impact on computer vision and NLP domains but has not been used in the area of recommender systems. Even though there has been a large body of research on generating recommendations based on modeling user-item interaction sequences, few of them attempt to represent and transfer these models for serving downstream tasks where only limited data exists. In this paper, we delve on the task of effectively learning a single user representation that can be applied to a diversity of tasks, from cross-domain recommendations to user profile predictions. Fine-tuning a large pre-trained network and adapting it to downstream tasks is an effective way to solve such tasks. However, fine-tuning is parameter inefficient considering that an entire model needs to be re-trained for every new task. To overcome this issue, we develop a parameter efficient transfer learning architecture, termed as PeterRec, which can be configured on-the-fly to various downstream tasks. Specifically, PeterRec allows the pre-trained parameters to remain unaltered during fine-tuning by injecting a series of re-learned neural networks, which are small but as expressive as learning the entire network. We perform extensive experimental ablation to show the effectiveness of the learned user representation in five downstream tasks. Moreover, we show that PeterRec performs efficient transfer learning in multiple domains, where it achieves comparable or sometimes better performance relative to fine-tuning the entire model parameters. Codes and datasets are available at https://github.com/fajieyuan/sigir2020_peterrec.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源