论文标题

部分可观测时空混沌系统的无模型预测

Sequential Recommendation with Causal Behavior Discovery

论文作者

Wang, Zhenlei, Chen, Xu, Zhou, Rui, Dai, Quanyu, Dong, Zhenhua, Wen, Ji-Rong

论文摘要

顺序建议的关键在于准确的项目相关建模。先前的模型根据项目共发生推断此类信息,这些信息可能无法捕获真正的因果关系,并影响建议性能和解释性。在本文中,我们为顺序推荐提供了一个新颖的因果发现模块,以捕获用户行为之间的因果关系。我们的总体想法首先假设有因果图的基础项目相关性,然后我们通过拟合真实的用户行为数据来共同学习与顺序推荐模型的因果图。更具体地说,为了满足因果关系的要求,因果图由可区分的定向无环约束正规化。考虑到推荐系统中的项目数量可能非常大,我们代表具有统一的潜在簇的不同项目,并且在集群级别定义了因果图,从而增强了模型的可扩展性和稳健性。此外,我们还提供了有关学习因果图的可识别性的理论分析。据我们所知,本文迈出了将顺序建议与因果发现结合的第一步。为了评估建议性能,我们使用不同的神经顺序体系结构实现了框架,并将它们与基于实际数据集的许多最新方法进行比较。经验研究表明,我们的模型平均可以在F1和NDCG上分别提高约7%和11%。为了评估模型的解释性,我们构建了一个新的数据集,其中包含人类标记的解释,以进行定量和定性分析。

The key of sequential recommendation lies in the accurate item correlation modeling. Previous models infer such information based on item co-occurrences, which may fail to capture the real causal relations, and impact the recommendation performance and explainability. In this paper, we equip sequential recommendation with a novel causal discovery module to capture causalities among user behaviors. Our general idea is firstly assuming a causal graph underlying item correlations, and then we learn the causal graph jointly with the sequential recommender model by fitting the real user behavior data. More specifically, in order to satisfy the causality requirement, the causal graph is regularized by a differentiable directed acyclic constraint. Considering that the number of items in recommender systems can be very large, we represent different items with a unified set of latent clusters, and the causal graph is defined on the cluster level, which enhances the model scalability and robustness. In addition, we provide theoretical analysis on the identifiability of the learned causal graph. To the best of our knowledge, this paper makes a first step towards combining sequential recommendation with causal discovery. For evaluating the recommendation performance, we implement our framework with different neural sequential architectures, and compare them with many state-of-the-art methods based on real-world datasets. Empirical studies manifest that our model can on average improve the performance by about 7% and 11% on f1 and NDCG, respectively. To evaluate the model explainability, we build a new dataset with human labeled explanations for both quantitative and qualitative analysis.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源