论文标题

交互式推荐系统通过知识图形增强的增强学习

Interactive Recommender System via Knowledge Graph-enhanced Reinforcement Learning

论文作者

Zhou, Sijin, Dai, Xinyi, Chen, Haokun, Zhang, Weinan, Ren, Kan, Tang, Ruiming, He, Xiuqiang, Yu, Yong

论文摘要

交互式推荐系统(IRS)由于其灵活的推荐策略和考虑最佳的长期用户体验而引起了极大的关注。为了应对动态的用户偏好并优化累积公用事业,研究人员将强化学习(RL)引入了IRS。但是,RL方法共享一个常见的样本效率问题,即需要大量的交互数据来培训有效的建议政策,这是由用户稀疏响应和由大量候选项目组成的大型动作空间引起的。此外,在在线环境中收集具有探索性策略的大量数据是不可行的,这可能会损害用户体验。在这项工作中,我们研究了利用知识图(KG)处理IRS的RL方法问题的潜力,该方法为建议决策提供了丰富的侧面信息。我们没有从头开始学习RL策略,而是利用从KG学到的项目相关性的先验知识来指导候选人选择更好的候选项目检索,(ii)丰富项目和用户状态的表示,以及(iii)在KG中宣传用户偏好,以应对kg的用户反馈的派发性。在两个现实世界数据集上进行了全面的实验,这证明了我们的方法的优势,并针对最先进的做法进行了重大改进。

Interactive recommender system (IRS) has drawn huge attention because of its flexible recommendation strategy and the consideration of optimal long-term user experiences. To deal with the dynamic user preference and optimize accumulative utilities, researchers have introduced reinforcement learning (RL) into IRS. However, RL methods share a common issue of sample efficiency, i.e., huge amount of interaction data is required to train an effective recommendation policy, which is caused by the sparse user responses and the large action space consisting of a large number of candidate items. Moreover, it is infeasible to collect much data with explorative policies in online environments, which will probably harm user experience. In this work, we investigate the potential of leveraging knowledge graph (KG) in dealing with these issues of RL methods for IRS, which provides rich side information for recommendation decision making. Instead of learning RL policies from scratch, we make use of the prior knowledge of the item correlation learned from KG to (i) guide the candidate selection for better candidate item retrieval, (ii) enrich the representation of items and user states, and (iii) propagate user preferences among the correlated items over KG to deal with the sparsity of user feedback. Comprehensive experiments have been conducted on two real-world datasets, which demonstrate the superiority of our approach with significant improvements against state-of-the-arts.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源